Hi there! It’s probably time to make this somewhat official:
Here is a selection of the most pressing “big ticket” animation related developments currently on my todo list. Do note that this is not an exhaustive list (for which there are many other items), but it does contain all the main things that I’m most aware of.
(This is cross-posted from my original post: http://aligorith.blogspot.co.nz/2015/03/animation-system-roadmap.html)
High Priority
NLA
* Local Strip Curves – Keyframing strip properties (e.g. time and influence) currently doesn’t update correctly. [2.75]
Quite frankly, I’m surprised the current situation seems to work as well as it has, because the original intention here (and only real way to solve it properly) is to have dedicated FCurves which get evaluated before the rest of the animation is handled.
I’ve got a branch with this functionality working already – all that’s missing is code to display those FCurves somewhere so that they can be edited (and without being confused for FCurves in the active actions instead). That said, the core parts of this functionality are now solid and back under control in the way it was originally intended.
I originally wanted to get this polished and into master for 2.74 – definitely before Gooseberry start trying to animate, as I know that previous open movie projects did end up using the NLA strip times for stuff (i.e. dragon wings when flying), and the inclusion of this change will be somewhat backwards incompatible (i.e. the data structures are all still there – nothing changed on that front, but there were some bugs in the old version which means that even putting aside the fact you can’t insert keyframes where they’re actually needed, the animations wouldn’t actually get evaluated correctly!).
On a related note – the bug report regarding the renaming NLA strips not updating the RNA Paths: that is a “won’t fix”, as that way of keyframing these properties (that is used in master) was never the correct solution. This fix will just simply blow it all away, so no point piling another hack-fix on top of it all.
* Reference/Rest Track and Animation Layers Support [2.76]
This one touches on two big issues. Firstly, there’s the bug where, if not all keyframed properties are affected by every strip (or at least set to some sane value by a “reference” strip), you will get incorrect poses when using renderfarms or jumping around the timeline in a non-linear way.
On another front, the keyframing on top of existing layers (i.e. “Animation Layers”) support doesn’t work well yet, because keyframing records the combined value of the stack + the delta-changes applied by the active action that you’re keying into. For this to work correctly, the contributions of the NLA stack must be able to be removed from the result, leaving only the delta changes, thus meaning that the new strip will be accumulated properly.
So, the current plan here is that an explicit “Reference Pose” track will get added to the bottom of NLA stacks. It will always be present, and should include every single property which gets animated in the NLA stack, along with what value(s) those properties should default to in the absence of any contributions from NLA strips.
Alongside this reference track, all the “NlaEvalChannels” will be permanently stored (during runtime only; they won’t get saved to the file) instead of being recreated from scratch each time. They will also get initialised from the Reference Track. Then, this allows the keyframing tools to quickly look up the NLA stack result when doing keyframing, thus avoiding the problems previously faced.
* A better way to retime a large number of strips [2.76/7]
It’s true that the current presentation of strips is not exactly the most compact of representations. To make it easier to retime a large number of strips (i.e. where you might want them to be staggered across a large number of objects, we may need to consider having something like a summary-track in the dopesheet. Failing that, we could just have an alternative display mode which compacts these down for this usecase.
Action Management [2.74, 2.75]
See the Action Management post. The priority of this ended up being bumped up, displacing the NLA fixes from 2.74 (i.e. Local Strip Keyframes) and 2.75 (i.e. Reference Track Support) back by 1-2 releases.
There are also a few related things which were not mentioned in that post (as they did not fit):
* Have some way of specifying which “level” the “Action Editor” mode works on.
Currently, it is strictly limited to the object-level animation of the active object. Nothing else. This may be a source of some of the confusion and myths out there… (Surely the fact that the icon for this mode uses the Object “cube” is a bit of a hint that something’s up here!)
* Utilities for switching between Dopesheet and NLA.
As mentioned in the Action Management post, there are some things which can be done to make the relationship between these closer, to make stashing and layering workflows nicer.
Also in question would be how to include the Graph Editor in there somehow too… (well, maybe not between the NLA, but at least with the Dopesheet)
* “Separate Curves” operator to split off FCurves into another action
The main point of this is to split off some unchanging bones from an action containing only moving parts. It also paves the way for other stuff like take an animation made for grouped objects back to working on individual objects.
Animation Editors
* Right-click menus in the Channels List for useful operations on those [2.75]
This should be a relatively simple and easy thing to do (especially if you know what to do). So, it should be easy to slot this in at some point.
* Properties Region for the Action Editor [2.76]
So, at some point recently, I realised that we probably need to give the Action Editor a dedicated properties region too to deal with things like groups and also the NLA/AnimData/libraries stuff. Creating the actual region is not really that difficult. Again it boils down to time to slot this in, and then figuring out what to put in there.
* Grease Pencil integration into normal Dopesheet [2.76]
As mentioned in the Grease Pencil roadmap, I’ve got some work in progress to include Grease Pencil sketch-frames in the normal dopesheet mode too. The problem is that this touches almost every action editor operator, which needs to be checked to make sure it doesn’t take the lazy road out by only catering for keyframes in an either/or situation. Scheduling this to minimise conflicts with other changes is the main issue here, as well as the simple fact that again, this is not “simple” work you can do when half-distracted by other stuff.
Bone Naming [2.77]
The current way that bones get named when they are created (i.e. by appending and incrementing the “.xyz” numbers after their names) is quite crappy, and ends up creating a lot of work if duplicating chains like fingers or limbs. That is because you now have to go through, removing these .xyz (or changing them back down to the .001 and .002 versions) before changing the action things which should change (i.e. Finger1.001.L should become Finger2.001.L instead of Finger1.004.L or Finger1.001.L.001).
Since different riggers have different conventions, and this functionality needs to work with the “auto-side” tool as well as just doing the right thing in general, my current idea here is to give each Armature Datablock a “Naming Pattern” settings block. This would allow riggers to specify how the different parts of each name behave.
For example, [Base Name][Chain Number %d][Segment Letter][Separator ‘.’][Side LetterUpper] would correspond to “Finger2a.L”. With this in place, the “duplicate” tool would know that if should increment the chain number/letter (if just a single chain, while perhaps preparing for flipping the entire side if it’s more of a tree), while leaving the segment alone. Or the “extrude” tool would know to increment the segment number/letter while leaving the chain number alone (and not creating any extra gunk on the end that needs to be cleaned up). The exact specifics though would need to be worked out to make this work well.
Drivers
* Build a dedicated “Safe Python Subset” expression engine for running standard driver expressions to avoid the AutoRun issues
I believe that the majority of driver expressions can be run without full Python interpreter support, and that the subset of Python needed to support the kinds of basic math equations that the majority of such driver expressions use is a very well defined/small set of things.
This set is small enough that we can in fact implement our own little engine for it, with the benefit that it could probably avoid most of the Python overheads as a result, while also being safe from the security risks of having a high-powered turing-complete interpreter powering it. Other benefits here are that this technique would not suffer from GIL issues (which will help in the new depsgraph; oddly, this hasn’t been a problem so far, but I’d be surprised if it doesn’t crop up its ugly head at the worst possible moment of production at some point).
In the case where it cannot in fact handle the expression, it can then just turf it over to the full Python interpreter instead. In such cases, the security limiting would still apply, as “there be dragons”. But, for the kinds of nice + simple driver expressions we expect/want people to use, this engine should be more than ample to cope.
So, what defines a “nice and simple” driver expression?
– The only functions which can be used are builtin math functions (and not any arbitrary user-defined ones in a script in the file; i.e. only things like sin, cos, abs, … would be allowed)
– The only variables/identifiers/input data it can use are the Driver Variables that are defined for that driver. Basically, what I’ve been insisting that people use when using drivers.
– The only “operators” allowed are the usual arithmetic operations: +, -, *, /, **, %
What makes a “bad” (or unsafe) driver expression?
– Anything that tries to access anything using any level of indirection. So, this rules out all the naughty “bpy.data[…]…” accesses and “bpy.context.blah” that people still try to use, despite now being blasted with warnings about it. This limitation is also in place for a good reason – these sorts of things are behind almost all the Python exploits I’ve seen discussed, and implementing such support would just complicate and bloat out little engine
– Anything that tries to do list/dictionary indexing, or uses lists/dictionaries. There aren’t many good reasons to be doing this (EDIT: perhaps randomly chosing an item from a set might count. In that case, maybe we should restrict these to being “single-level” indexing instead?).
– Anything that calls out to a user-defined function elsewhere. This is inherent risk here, in that that code could do literally anything
– Expressions which try to import any other modules, or load files, or crazy stuff like that. There is no excuse… Those should just be red-flagged whatever the backend involved, and/or nuked on the spot when we detect this.
* A modal “eyedropper” tool to set up common “garden variety” 1-1 drivers
With the introduction of the eyedropped tools to find datablocks and other stuff, a precedent has been set in our UI, and it should now be safe to include similar things for adding a driver between two properties. There are of course some complications which arise from the operator/UI code mechanics last time I tried this, but putting this in place should make it easier for most cases to be done.
* Support for non-numeric properties
Back when I initially set up the animation system, I couldn’t figure out what to do with things like strings and pointers to coerce them into a form that could work with animation curves. Even now, I’m not sure how this could be done. That said, while writing this, I had the though that perhaps we could just use the same technique used for Grease Pencil frames?
Constraints
* Rotation and Scale Handling
Instead of trying to infer the rotation and scale from the 4×4 matrices (and failing), we would instead pass down “reference rotation” and “reference scale” values alongside the 4×4 matrix during the evaluation process. Anytime anything needs to extract a rotation or scale from the matrix, it has to adjust that to match the reference transforms (i.e. for rotations, this does the whole “make compatible euler” stuff to get them up to the right cycle, while for scale, this just means setting the signs of the scale factors). If however the rotation/scale gets changed by the constraint, it must also update those to be whatever it is basing its stuff from.
These measures should be enough to combat the limitations currently faced with constraints. Will it result in really ugly code? Hell yeah! Will it break stuff? Quite possibly. Will it make it harder to implement any constraints going forth? Absolutely. But will it work for users? I hope so!
Rigging
It’s probably time that we got a “Rigging Dashboard” or similar…
Perhaps the hardest thing in trying to track down issues in the rigs being put out by guys like JP and cessen these days are that they are so complex (with multiple layers of helper bones + constraints + parenting + drivers scattered all over) to figure out where exactly to start, or which set of rigging components interact to create a particular result.
Simply saying “nodify everything” doesn’t work either. Yes, it’s all in one place now, but then you’ve got the problem of a giant honking graph that isn’t particularly nice to navigate (large graph navigation in and of itself is another interesting topic for another time and date).
Key things that we can get from having such a dashboard are:
1) Identifying cycles easier, and being able to fix them
2) Identifying dead/broken drivers/constraints
3) Isolating particular control chains to inspect them, with everything needed presented in one place (i.e. on a well designed “workbench” for this stuff)
4) Performance analysis tools to figure out which parts of your rig are slow, so that you can look into fixing that.
Medium Priority
NLA
* A better way of flattening the stack, with fewer keyframes created
In many cases, it is possible to flatten the NLA without baking out each frame. This only really applies when there are no overlaps, where the keyframes can simply be transposed “as is”. When they do interact though, there may be possibilities to combine these in a smarter way. In the worst case, we can just combine by baking.
* Return of special handling for Quaternions?
I’m currently pondering whether we’ll need to reinstate special handling for quaternion properties, to keep things sane when blending.
* Unit tests for the whole time-mapping math
I’ve been meaning to do this, but I haven’t been able to get the gtests framework to work with my build system yet… If there ever wee a model example of where these things come in handy, it is this!
Animation Editors
* Expose the Animation Channel Filtering API to Python
Every time I see the addons that someone has written for dealing with animation data, I’m admittedly a bit saddened that they do things like explicitly digging into the active object only, and probably only caring about certain properties in there. Let’s just say, “been there done that”… that was what was done in the old 2.42/3 code, before I cleaned it up around 2.43/2.44, as it was starting to become such a pain to maintain it all (i.e. each time a new toggle or datatype was added, ALL the tools needed to be recoded).
These days, all the animation editors do in fact use a nice C API for all things channels-related. Some of it pre-dates the RNA system, so it could be said that there are some overlaps. Then again, this one is specialised for writing animation tools and drawing animation editors, while RNA is generic data access – no comparison basically.
So, this will happen at some point, but it’s not really an urgent/blocking issue for anything AFAIK.
* To support the filtering API, we need a way of setting up or supplying some more general filtering settings that can be used everywhere where there aren’t any the dopesheet filtering options already
The main reason why all the animation editor operators refuse to work outside of those editors is that they require the dopesheet filtering options (i.e. those toggles on the header for each datablock, and other things) to control what they are able to see and affect. If we have some way of passing such data to operators which need it in other contexts (as a fallback), this opens the way up for stuff like being able to edit stuff in the timeline.
As you’ll hopefully be well aware, I’m extremely wary of any requests to add editing functionality to the timeline. On day one, it’ll just be “can we click to select keyframes, and then move them around”, and then before long, it’s “can we apply interpolation/extrapolation/handle types/etc. etc.” As a result, I do not consider it viable to specifically add any editing functionality there. If there is editing functionality for the timeline, it’ll have to be borrowed from elsewhere!
Action Editor/Graph Editor
* Add/Remove Time
Personally I don’t understand the appeal of this request (maybe it’s a Maya thing), but nonetheless, it’s been on my radar/list as something that can be done. The only question is this: is it expected that keyframes should be added to enact a hold when this happens, or is this simply expanding and contracting the space between keyframes.
* Make breakdown keyframes move relative to the main keyframes
In general, this is simple, up until the keyframes start moving over each other. At that point, it’s not clear how to get ourselves out of that pickle…
Small FCurve/Driver/etc. Tweaks
* Copy Driver Variables
* Operators to remove all FModifiers
Motion Capture Data
* A better tool for simplifying dense motion curves
I’ve been helping a fellow kiwi work on getting his curve simplifying algorithm into Blender. So far, its main weakness is that it is quite slow (it runs in exponential time, which sucks on longer timelines) but has guarantees of “optimal” behaviour. We also need to find some way to estimate the optimal parameters, so that users don’t have to spend a lot of time testing different combinations (why is not going to be very nice, given the non-interactive nature of this).
Feel free to try compiling this and give it a good test on a larger number of files and let us know how you go!
* Editing tools for FSamples
FSamples were designed explicitly for the problem of tackling motion capture data, and should be more suited to this than the heavier keyframes.
Keying Sets
* Better reporting of errors
The somewhat vague “Invalid context” error for Keying Sets comes about because there isn’t a nice way to pipe more diagnostic information in and out of the Keying Sets callbacks which can provide us with that information. It’s a relatively small change, but may be better with
Pose Libraries
* Internal code cleanups to split out the Pose Library API from the Pose Library operators
These used to be able to serve both purposes, but the 2.5 conversion meant that they were quickly converted over to opertator-only to save time. But, this is becoming a bottleneck for other stuff
* Provide Outliner support for Pose Library ops
There’s a patch in the tracker, but this went about this in the wrong way (i.e. by duplicating the code into the outliner). If we get that issue out of the way, this is relatively trivial
* Pose Blending
Perhaps the biggest upgrade that can be made is to retrofit a different way of applying the poses, to be one which can blend between the values in the action and the current values on the rig. Such functionality does somewhat exist already (for the Pose Sliding tools), but we would need to adapt/duplicate this to get the desired functionality. More investigation needed, but it will happen eventually.
* Store thumbnails for Poses + Use the popup gallery (i.e. used for brushes) to for selecting poses
I didn’t originally do this, as at the time I thought that these sorts of grids weren’t terribly effective (I’ve since come around on this, after reading more about this stuff) and that it would be much nicer if we could actually preview how the pose would apply in 3D to better evaluate how well it fits for the current pose (than if you only had a 2D image to work off). The original intent was also to have a fancy 3D gallery, where scrolling through the gallery would swing/slide the alternatively posed meshes in from the sides.
Knowing what I know now, I think it’s time we used such a grid as one of the way to interact with this tool. Probably the best way would be to make it possible to attach arbitrary image datablocks to Pose Markers (allowing for example the ability to write custom annotations – i.e. what phenoms a mouth space refers to), and to provide some operators for creating these thumbnails from the viewport (i.e. by drawing a region to use).
Fun/Useful but Technically Difficult
There are also a bunch of requests I’d like to indulge, and indeed I’ve wanted to work on them for years. However, these also come with a non-insignificant amount of baggage which means that they’re unlikely to show up soon.
Onionskinning of Meshes
Truth be told, I wanted to do this back in 2010, around the time I first got my hands on a copy of Richard William’s book. The problem though was and remains that of maintaining adequate viewport/update performance.
The most expensive part of the problem is that we need to have the depsgraph (working on local copies of data, and in a separate thread) stuff in place before we can consider implementing this. Even then, we’ll also need to include some point caching stuff (e.g. Alembic) to get sufficient performance to consider this seriously.
Editable Motion Paths
This one actually falls into the “even harder” basket, as it actually involves 3-different “hard” problems:
1) Improved depsgraph so that we can have selective updates of only the stuff that changes, and also notify all the relationships appropriately
2) Solving the IK problem (i.e. changed spline points -> changed joint positions -> local-space transform properties with everything applied so that it works when propagated through the constraints ok). I tried solving this particular problem 3 years ago, and ran into many different little quirky corner cases where it would randomly bug/spazz out, flipping and popping, or simply not going where it needs to go because the constraints exhibit non-linear behaviour and interpret the results differently. This particular problem is one which affects all the other fun techniques I’d like to use for posing stuff, so we may have to solve this once and for all with an official API for doing this. (And judging from the problems faced by the authors of various addons – including the current editable motion paths addon, and also the even greater difficulties faced by the author of the Animat on-mesh tools, it is very much a tricky beast to tame)
3) Solving the UI issues with providing widgets for doing this.
Next-Generation Posing Tools
Finally we get to this one. Truth be told, this is the project I’ve actually been itching to work on for the past 3 years, but have had to put off for various reasons (i.e. to work on critical infrastructure fixes and also for uni work). It is also somewhat dependent on being able to solve the IK problem here (which is a recurring source of grief if we don’t do it right).
If you dig around hard enough, you can probably guess what some of these are (from demos I’ve posted and also things I written in various places). The short description though is that, if this finally works in the way I intend, we’ll finally have an interface that lets us capture the effortless flow, elegance, and power of traditional animating greats like Glen Keane or Eric Goldberg – for having a computer interface that allows that kind of fluid interaction is one my greatest research interests.
Closing Words
Looking through this list, it looks like we’ve got enough here for at least another 2-3 years of fun times :D
greatness!
Especially the last few visions(wish scene\Vision of hope).
The above content is expressed through translation software.
This is the best implementation of an editable motion path that I’ve come across so far.
https://www.toolchefs.com/?portfolio=motion-path
It’s fast, and give excellent visual indicators to the animator.
You can also switch it between world space, and camera space.
What I’d really like to see is to have Blender Renderer’s materials able to handle custom written GLSL shaders. I use blender to create assets for video games, and it would be really useful to model in Blender using the GLSL shaders that I’m going to use in the game.
That Rigging Dashboard sounds awesome.
And good to see that Constraint scaling issues will be addressed as well.
Thanks Aligorith & Blender Team.
What about center the 3D cursor with one key?
Concerning the GIL… I really don’t think this will ever be a problem, since Python threading is currently impossible in Blender and will stay that way for the foreseeable future.
Have you thought about using sympy expressions? These can be compiled to various formats, and you can reason about their safety. Also automatic integration/differentiation. Would probably be too much of an integration project, though.
Another solution short of writing your own Python parser would be leveraging Python’s AST module, and then manually validating that all called functions are safe.
Just Finished reading the post:
Lots of good stuff here!
Storing thumbnails for pose librarys (…) YES PLEASE!!! using a brushlike ui for that is a great idea!
If by Editable Motion paths you mean something like After Effects, that you can edit on the 3d view without being on the specific frame… That was someting I put on my list also :D.
Thank you for all your work, for all those years!
Sorry I got distracted and sent the message. My question is: How can that list help the development… Is there somewhere to send it?
Thank you guys!
I´ll take time to read all this post.
When we were making our short film, Pique Tico, i thought that it would be good to take notes about improvements of the software. So I could help the development of it.
I have a long list, some on the animation tools, for instance some like:
Find/show locked channels (It may be like a lock or something like that on the menu bar, something like the ghost or the pointer arrow)
A 3d view mode to watch only silouetes of the characters.
Shape key locks, to make shape keys uneditable, sometimes by mistake you screw one or more shape keys. This stack can be hightly improved, I don´t know if it is hard to do but it also could work by bins or folders to organize the high amount of shape keys we do.
Wave form in the timeline. To get feedback of our animation lipsync.
Clear unused or broken drivers.
Separate the gl render output configuration from the render, so we dont have to change paths and formats, sometimes you are correcting animation and need both, the playblast and the render.
And lots more. I´m NOT asking anything of this, I just want to show you that I have a list that came out by requirements of the work and not by whims.
This is awsome, finally!
Please watch this youtubeREMOVEcom/watch?REMOVEv=1fFv9znBZOo
stonefield.cocolog-nifty.REMOVEcom/higurashi/2014/08/bone-modeling-1.REMOVEhtml
Incredibly useful addon for work with bones(its translated on many languages including English). Could you make those features in Blender(not via addon) as well? Pleeeease.
Joshua!
This sounds so great and exciting. How can animators help you with this? do you have an base of testers, or are more people needed. I’m especially excited about the next gen posing tools, but I see that it’s further down the lane.
Cheers
I second move keys in timeline.
also something that bugs me greatly is such a simple thing yet puts me off animating. say isolate the x location curve in the graph editor then i move a curve, i don’t like the move, then press ctrl z to undo the move, it turns all the curves on again.
hope those get sorted
Moving keys IN THE TIMELINE PLEASE !!!
See above :)
It’s not going to be something that will be specially done, as it’ll really be just the tip of a (very large) iceberg.
Speaking of animation, these rigs selection tools are awesome: https://www.youtube.com/watch?v=liqn03OKQBg
I know you have to much to consider right now, but this does make animation look easier!
Sorry, if I’m being off topic or redundant.
While I haven’t actually mentioned those here, yes these are an interesting direction for where the industry at large should be heading.
Regarding on-mesh selection of controls:
1) From the looks of things, what is probably happening is simply that they’ve set things up so that all the fine-tuning controls are simply buried slightly below the surface of the mesh, making them invisible unless specifically shown. Since there are no other controls nearby, selection uses works as intended. I’ve used this approach myself in the past when experimenting with better ways to build rigs.
2) It is however possible that they just did what Pixar did, and implemented some genuine “mesh zones” technique (similar to Fast Animation Surface Tookit, the unreleased Animat addon, or the WIP stuff Antony was playing with in the Gooseberry/Widgets branches).
Regarding on-mesh highlighting: I’ve done experiments before where I used the Mask Modifier to do this (that’s actually what it’s original purpose was when I built it ;) – use the “Armature” mode instead of the default “Vertex Group’.
Then, there’s also the possibility of these things getting built as addons. The Animat one is a good example, though I too would like to know what’s going on there (*friendly poke*). From what I gather though, it too is one of the tools which ran into trouble trying to support nicer direct-editing of stuff where the IK constraints/chains and/or constraints are involved/
Finally, the last item on the list there really does seek to make the issues of selecting stuff a bit less relevant (though previous attempts have convinced me that it would be wise to at least allow the use of selection to control/limit the region of controls affected)
I didn’t notice something that is important to me as a character animator : A motion path relative to the camera ! At the moment, the motion path works fine to track anobject or bone in the screen as long as the camera doesn’t move. But if you have a moving camera, the motion path is just useless to track your animation (on my last project, I had a camera following a bird for exemple)
Also too bad your poselib improvments fell at the bottom at the list (thumbnails and mixer !!)
Hope you get the chance as much as possible from your list !
Regarding these relative motionpaths:
I’m not sure that I get what you’re talking about. How would the placement of the camera affect this?
Regarding the pose library stuff:
Oh! Haha, I didn’t actually intend for it to be at the very bottom of that section! The main reason it’s not in the “high priority” section is that it is very much possible to still animate/work without this working, whereas those in the things like NLA evaluation or constraint matrix decomposition are very much things where working around them becomes tricky.
Could you please allow usage of “-” or “_” in the bone naming, besides “.” ?
“.” is not friendly for some game engines when exporting skeletal meshes.
i was wondering since 2.5 why the “-” was remove from bones autorename?, it was really usefull to have instead only the “.”
That’s a bit surprising… It’s possible that this happened as part of one round of refactoring, when we unified all the different autonaming algorithms. I can’t remember who worked on that exactly, though it would’ve been either Campbell or me.
The proposed bone naming template would allow you to do just that. So, in the example above, you could use:
[Base Name][Chain Number %d][Segment Letter][Separator ‘-‘][Side LetterUpper]
OR
[Base Name][Chain Number %d][Segment Letter][Separator ‘_’][Side LetterUpper]
instead (i.e. Note what happens with the [Separator ‘x’] bit :)
@Ross Franks
Thats scripting area, there are already some old scripts for older versions of blender, that (kinda) roboticaly walk. Those scripts not always work on recent versions, and the plans above, i think we would need a new script for it, not sure what directions they go with the python api on animation.
For reference, the API changes above would not have much bearing on those. The changes mentioned above mainly apply for writing tools for editing keyframes (e.g. the NLA baking tool, or perhaps some kind of add/remove time script in the meantime before I get to it).
Regarding autowalkers: Again, these are highly dependent on rig setup, and would only be suitable for background characters. That said, there’s nothing to stop anyone from writing one of these scripts now (i.e. I know Bassam has done a few, and then there are the old ones like the Walk-o-matic back in the day :)
A lot of good stuff here I still have to absorb it all. I need to write a bug/feature request for something that’s been bugging me but here it is anyway. Keying Sets really need to be written so that they respect locked channels. It seems pretty broken the way they work now.
The “Whole Character” keying set actually does this already.
For the other ones, it’s just a case of modifying the “generator” callbacks to include provisions for this.
From what I’ve seen the character keying set does not work as expected. As far as I can tell, unless you create a set that specifically excludes certain channels, a keying set will add keys to locked channels. Am I wrong about this?
Awesome work!
Sorry I havent read the whole post, but I hope in the future I’d like to see something like 3DS Max’s footstep system, where you plant footsteps and the character follows them (even up stairs etc)
Currently there are no plans to have develop any new builtin walkcycle/footsteps tools.
Previously we did have a few which were in various states of brokenness, and were tightly coupled/reliant on certain styles of rigging setups. For this reason, I think it would be better if for the time being, such functionality is developed as addons (which can be tailored to the rigs involved as necessary).
If however a sufficiently good/general solution does come up though, I’d gladly looking into building it in Blender proper. :)
May the Force be with you, Joshua!
Ref. PyDrivers, PLEASE reinstate the “.self” that was available up to Blender 2.49!
Ref. PyDriver restrictions, I have nothing against sandboxes, but, please:
let the knowledgeable user break out of the sandbox after some warnings, yes?
Where and why is this useful?
Do you expect self to point to the datablock that the driver is attached to, or do you expect this to be the struct that the driven property lives on (e.g. the bone vs the object)?
“.self” as in a ShapeKey value expressed as
min( pi/6 , mathutils.Vector((0.0 , 0.0 , 1.0)).angle( self.matrix_world.to_3x3() * mathutils.Vector((0.0 , 0.0 , 1.0)) ) ) / (pi/6)
i.e. the value of a ShapeKey dependent on the angle between the ShapeKey owner’s local Z
(hence, the need for self.matrix_world) and World Z.
An expression similar to the above (with math.pi, etc.) does work, if run from Text Editor.
In order to prevent spam, comments are closed 7 days after the post is published. Feel free to continue the conversation on the forums.