January 8th, 2014 at 5:00pm PDT, Agenda and Minutes

From Web3D.org
Revision as of 02:16, 9 January 2014 by Brutzman (Talk | contribs) (2. Relationship to current and future H-Anim)

Jump to: navigation, search

Dick Puk, Don Brutzman, William Glascoe, Myeong Won Lee, Joe Williams, and Prof. Jung-Ju Choi attended.

1. Facial animation using current H-Anim

Jung-Ju explained his proposal for facial animation. Joe Williams answered Jung-Ju's questions about his proposal. Check out MPEG-4 Facial Animation Parameter overview at the following URL.

Everyone has to be very careful not to infringe on MPEG4 FAP Intellectual Property Rights in facial animation

  • Virtual Humans by Prof Nadia Magnenat Thalmann is another reference.

2. Relationship to current and future H-Anim

Don explains how feature points are animated (via H-AnimSites) and how a comparison (coordinate interpolator of a mesh to see if a displacer is needed).

A prerequisite for coordinate interpolator is that every coordinate in the mesh must be animated.

  • In X3D, a CoordinateInterpolator can animate a mesh. This is an interesting alternative that can be compared in a test.
  • The red dots in the figure correspond to feature points
  • The number of black vertices might vary depending on an implementation
  • The most interesting choice for an H-Anim perspective might be choice of those feature points
    • H-Anim has about about 8 (or 9) feature points for face. See H-Anim Specification Figure B.1 and feature points 1..8, which are listed in Table B.1.
    • MPEG4 has about 50 feature points for face, but there are patents restricting their specific use in an open standard

So our first main design decision for H-Anim might be whether to 1. pre-specify a set of feature points for common use (and re-use) 2. or allow an author to define them (which might not be valuable, they can define geometry anyway

(I'm guessing we'd want option #1, but it is possible that we might find a way to do something like option #2.)

There are also a finite number of human facial expressions

  • there are multiple models published about this (see SIGGRAPH papers)
  • each expression has a name
  • there is a set of feature points common to these

So our next next decision might be 3. to define a set of sites, likely as a superset of the existing 8 points in H-Anim 4. to define a set of expressions, each named, each with a finite number of the feature points

Possibility: if we added a little more to Jung-Ju's point paper (perhaps steps 1 3 and 4 above) then we could write "expression animators" that describe the end locations of each feature point.

Joe's comments

A displacer's advantage is that you specify only the end value, not the initial value. A coordinate interpolator


There are animations using displacers. Joe recognizes this proposal is targeting an end point with a weight. In a way it works as an interpolator. Where do you get the source animations for facial expressions?

Level of Detail perspective (JW)...

3. Standardization item for facial animation

4. NWIP discussion

William mentioned collaborating with the Parametric Human Project as they amass microCT data of human skeletons. See a paper describing the project scope and goals.

5. Scheduling next meeting

Next meeting will be held in person, co-located at the meetings below, as well as via Web3D teleconference. January 20th (Monday) at 4:00pm PST (21th, 9:00am KST)


Wed3D, SC24 WG9, AR JAhG Meetings, January 20-23, The K Seoul Hotel, Seoul Korea Meeting Place: Heagem B, 3rd floor, The K Seoul Hotel (http://www.thek-hotel.co.kr/e_s! eoul/main.asp)

(OLD name: Seoul KyoYuk-MunHwa-HoiKwan)