Loading...
Logo
Processing Forum
Hi,

I'm trying to create a project where the users motion is detected and drawn on the screen. I have found many very helpful resources on this forum and at openProcessing.org, specifically the optical flow project by Hidetoshi Shimodaira. Thanks very much for such a valuable resource. Below is the code that I have so far. There are two things that I would like to change if possible and am looking for some suggestions. I like how the vectors show the direction of movement but ideally I would really like them to be displayed in the color of the corresponding pixel in the camera feed. So that if, for example, someone has a red shirt on and moves in front of the camera the vectors would show up red where they had moved. If that makes sense? And finally I would like to add a blur to the final image. Can I do that? Please excuse the excessive amount of //comments in my code, I'm really new to this and it helps me keep track of what I'm looking at. :)
Copy code




  1. int wscreen=640;//app window width
    int hscreen=480;//app window height
    int gs=5; //number of pixels per grid step
    float predsec=1.0; //larger number makes longer vector line

    import processing.opengl.*;
    import processing.video.*;//initialize camera
    Capture video;//capture video image
    int fps = 30; //frames per second
    color[] vline;


    //parameters for the grid
    int as=gs*2; //area of window for averaging pixels
    int gw=wscreen/gs;//width of screen/number of grid
    int gh=hscreen/gs;//height of screen/number of grid
    int gs2=gs/2;//grid/2
    float df=predsec*fps;//predsec*frames per second

    //regression vectors
    float[] fx, fy, ft;
    int fm=3*9; // length of the vectors

    // regularization term for regression
    float fc=pow(10,8); // larger values for noisy video

    // smoothing parameters
    float wflow=0.04; // smaller value for longer lasting vectors


    boolean flagmirror=true; // mirroring image
    boolean flagflow=true; // draw opticalflow vectors
    boolean flagimage=false; // show video image
    boolean flagseg=false; // segmentation of moving objects

    // internally used variables
    float ar,ag,ab; // used as return value of pixave
    float[] dtr, dtg, dtb; // differentiation by t (red,gree,blue)
    float[] dxr, dxg, dxb; // differentiation by x (red,gree,blue)
    float[] dyr, dyg, dyb; // differentiation by y (red,gree,blue)
    float[] par, pag, pab; // averaged grid values (red,gree,blue)
    float[] flowx, flowy; // computed optical flow
    float[] sflowx, sflowy; // slowly changing version of the flow
    int clockNow,clockPrev, clockDiff; // for timing check

    void setup() {
      // screen and video
      size(wscreen, hscreen, P2D);
      video = new Capture(this, wscreen, hscreen, fps);

      rectMode(CENTER);//setting region to captue pixels from

      // arrays
      par = new float[gw*gh];
      pag = new float[gw*gh];
      pab = new float[gw*gh];
      dtr = new float[gw*gh];
      dtg = new float[gw*gh];
      dtb = new float[gw*gh];
      dxr = new float[gw*gh];
      dxg = new float[gw*gh];
      dxb = new float[gw*gh];
      dyr = new float[gw*gh];
      dyg = new float[gw*gh];
      dyb = new float[gw*gh];
      flowx = new float[gw*gh];
      flowy = new float[gw*gh];
      sflowx = new float[gw*gh];
      sflowy = new float[gw*gh];
      fx = new float[fm];
      fy = new float[fm];
      ft = new float[fm];
      vline = new color[wscreen];
    }
    // calculate average pixel value (r,g,b) for rectangle region
    void pixave(int x1, int y1, int x2, int y2) {
      float sumr,sumg,sumb;
      color pix;
      int r,g,b;
      int n;

      if(x1<0) x1=0;
      if(x2>=wscreen) x2=wscreen-1;
      if(y1<0) y1=0;
      if(y2>=hscreen) y2=hscreen-1;

      sumr=sumg=sumb=0.0;
      for(int y=y1; y<=y2; y++) {
        for(int i=wscreen*y+x1; i<=wscreen*y+x2; i++) {
          pix=video.pixels[i];
          b=pix & 0xFF; // blue
          pix = pix >> 8;
          g=pix & 0xFF; // green
          pix = pix >> 8;
          r=pix & 0xFF; // red
          // averaging the values
          sumr += r;
          sumg += g;
          sumb += b;
        }
      }
      n = (x2-x1+1)*(y2-y1+1); // number of pixels
      // the results are stored in static variables
      ar = sumr/n;
      ag=sumg/n;
      ab=sumb/n;
    }

    // extract values from 9 neighbour grids
    void getnext9(float x[], float y[], int i, int j) {
      y[j+0] = x[i+0];
      y[j+1] = x[i-1];
      y[j+2] = x[i+1];
      y[j+3] = x[i-gw];
      y[j+4] = x[i+gw];
      y[j+5] = x[i-gw-1];
      y[j+6] = x[i-gw+1];
      y[j+7] = x[i+gw-1];
      y[j+8] = x[i+gw+1];
    }

    // solve optical flow by least squares (regression analysis)
    void solveflow(int ig) {
      float xx, xy, yy, xt, yt;
      float a,u,v,w;

      // prepare covariances
      xx=xy=yy=xt=yt=0.0;
      for(int i=0;i<fm;i++) {
        xx += fx[i]*fx[i];
        xy += fx[i]*fy[i];
        yy += fy[i]*fy[i];
        xt += fx[i]*ft[i];
        yt += fy[i]*ft[i];
      }

      // least squares computation
      a = xx*yy - xy*xy + fc; // fc is for stable computation
      u = yy*xt - xy*yt; // x direction
      v = xx*yt - xy*xt; // y direction

      // write back
      flowx[ig] = -2*gs*u/a; // optical flow x (pixel per frame)
      flowy[ig] = -2*gs*v/a; // optical flow y (pixel per frame)
    }

    void draw() {
      if(video.available()) {
        // video capture
        video.read();
    video.loadPixels();
        // mirror
        if(flagmirror) {
          for(int y=0;y<hscreen;y++) {
            int ig=y*wscreen;
            for(int x=0; x<wscreen; x++)
              vline[x] = video.pixels[ig+x];
            for(int x=0; x<wscreen; x++)
              video.pixels[ig+x]=vline[wscreen-1-x];
          }
        }

        // draw image
        if(flagimage) set(0,0,video);
        else background(255);

        // 1st sweep : differentiation by time
        for(int ix=0;ix<gw;ix++) {
          int x0=ix*gs+gs2;
          for(int iy=0;iy<gh;iy++) {
            int y0=iy*gs+gs2;
            int ig=iy*gw+ix;
            // compute average pixel at (x0,y0)
            pixave(x0-as,y0-as,x0+as,y0+as);
            // compute time difference
            dtr[ig] = ar-par[ig]; // red
            dtg[ig] = ag-pag[ig]; // green
            dtb[ig] = ab-pab[ig]; // blue
            // save the pixel
            par[ig]=ar;
            pag[ig]=ag;
            pab[ig]=ab;
          }
        }

        // 2nd sweep : differentiations by x and y
        for(int ix=1;ix<gw-1;ix++) {
          for(int iy=1;iy<gh-1;iy++) {
            int ig=iy*gw+ix;
            // compute x difference
            dxr[ig] = par[ig+1]-par[ig-1]; // red
            dxg[ig] = pag[ig+1]-pag[ig-1]; // green
            dxb[ig] = pab[ig+1]-pab[ig-1]; // blue
            // compute y difference
            dyr[ig] = par[ig+gw]-par[ig-gw]; // red
            dyg[ig] = pag[ig+gw]-pag[ig-gw]; // green
            dyb[ig] = pab[ig+gw]-pab[ig-gw]; // blue
          }
        }

        // 3rd sweep : solving optical flow
        for(int ix=1;ix<gw-1;ix++) {
          int x0=ix*gs+gs2;
          for(int iy=1;iy<gh-1;iy++) {
            int y0=iy*gs+gs2;
            int ig=iy*gw+ix;

            // prepare vectors fx, fy, ft
            getnext9(dxr,fx,ig,0); // dx red
            getnext9(dxg,fx,ig,9); // dx green
            getnext9(dxb,fx,ig,18);// dx blue
            getnext9(dyr,fy,ig,0); // dy red
            getnext9(dyg,fy,ig,9); // dy green
            getnext9(dyb,fy,ig,18);// dy blue
            getnext9(dtr,ft,ig,0); // dt red
            getnext9(dtg,ft,ig,9); // dt green
            getnext9(dtb,ft,ig,18);// dt blue

            // solve for (flowx, flowy) such that
            // fx flowx + fy flowy + ft = 0
            solveflow(ig);

            // smoothing
            sflowx[ig]+=(flowx[ig]-sflowx[ig])*wflow;
            sflowy[ig]+=(flowy[ig]-sflowy[ig])*wflow;
          }
        }


        // 4th sweep : draw the flow
        if(flagseg) {
          noStroke();
          fill(0);
          for(int ix=0;ix<gw;ix++) {
            int x0=ix*gs+gs2;
            for(int iy=0;iy<gh;iy++) {
              int y0=iy*gs+gs2;
              int ig=iy*gw+ix;

              float u=df*sflowx[ig];
              float v=df*sflowy[ig];


              float a=sqrt(u*u+v*v);
              if(a<2.0) rect(x0,y0,gs,gs);
            }
          }
        }

        // 5th sweep : draw the flow
        if(flagflow) {
          for(int ix=0;ix<gw;ix++) {
            int x0=ix*gs+gs2;
            for(int iy=0;iy<gh;iy++) {
              int y0=iy*gs+gs2;
              int ig=iy*gw+ix;

              float u=df*sflowx[ig];
              float v=df*sflowy[ig];

              // draw the line segments for optical flow
              float a=sqrt(u*u+v*v);
              if(a>=2.0) { // draw only if the length >=2.0
                float r=0.5*(1.0+u/(a+0.1));
                float g=0.5*(1.0+v/(a+0.1));
                float b=0.5*(2.0-(r+g));

                //  stroke(r,g,b);
                stroke(255*r, 255*b, 255*g);//color of line *this is what I want to change but I don't know what I should                                                                                 write to get the pixels of the camera input?
                line(x0,y0,x0+u,y0+v);//location of line beginning and end
               
         
              }
            }
          }
        }
      }
    }


Any assistance or tips towards potentially helpful tutorials would be greatly appreciated.

THANKS!!

Replies(1)

I like you'r work! 

I'm new in processing and just got my first program running properly. I wonder if it is possible to recognise if either green or red vectors are dominating in you'r program?

I look forward to your hear from you :)