Blender .fbx -> Spark AR Studio scaling issue for skeleton animations - blender

I'm trying to create a character with skeleton animation in Blender to bring into Spark AR Studio. In Spark I want to use the baked animation. The .fbx brings the model and skeleton into Spark's scene just fine, until a new animation controller is selected via the object's inspector window and the animation is selected for use.
At that point, the Empty object named "Armature" is scaled to 100 instead of 1 and cannot be changed.
As a workaround, the Skeleton child object named "skeleton" can be scaled to 0.01. In Blender, I tried changing the scene's units and made sure the object's scales were all applied. Nothing is scaled to 100, everything is scaled to 1.
Since the object from the .fbx imports into Spark with correct scaling, I expect the animation to maintain that, but once the animation is selected the scale jumps from 1 to 100.

Put you animated object inside some nullObject and then scale down not animated but nullObject. Hope it is clear.

Related

Blender glft mirrors armature after exporting model

I have gltf model with rig and animation sequences (with .bin), exported from video game League of Legends. After mport into Blender, everything looks fine (almost, but that's not important for rig), but after exporting into other format (I export as .smd to use in Source Filmmaker, but also tried othe formats) and importing into Blender exported model armature mirrors along global Y. Here's how animation has to look
And here is what comes after export-import:
import mesh with armature, but without animation
and what comes with adding animation
(for this after previous screenshot I flipped mesh in edit mode along Y global).
I tried to mirror before/after export-import both armature and mesh. Nothing helped before. On last screenshot everything is flipped along global X. One day I fixed it, but her weapon (circle thing) is always behind her after export-import when it must be ahead of her.

blender render object from multiple angles

I have an object in blender + an HDRI background / environment map. I am using cycles to render the object and I have Blender 2.8.
I would like to take multiple pictures of the rendered object (with its background) so that I end up with multiple views of the object (say, about 5/10).
I have seen some posts out there but they're not quite what I want because they just render in solid mode whereas I actually want the whole render.
I am a newbie with blender and I don't even know where to start with this. Thank you
You can just render multiple pictures by rendering, saving the image then moving the camera and repeating the process again. Or, you could render an animation with the camera moving to different angle and setting the output to an image format.

Blender border render internals

I would like to know how Blender's border render works internally. How can Blender compute lights if it has not information about the lights in the tiles he won't render? I have not found any reference (source code excluded) on how this feature of blender works. Can somebody explain it (or give me some reference)?
The render border setting only alters what part of the image is rendered, it does not alter what data is sent to the render engine to generate the image.
You can test this by placing an object with a reflective surface in front of the camera and another object behind the camera, the object behind the camera will show in the reflection. The border setting doesn't change the reflection in the object, it only changes what part of the image is rendered.
Rendering an image starts at the pixel that will be visible in the final image and sends a "ray" into the scene to determine what colour the specific pixel will be. Each ray will bounce around in the scene from object to object to light source based on render settings to calculate the final result. While the render border will reduce the pixels used as the starting point for each ray, it does not reduce the objects or lights in the scene that each ray may come into contact with. Each ray going through the scene will see every visible object and light in the scene that can influence the final result for each pixel.
This conference video explains ray types and might give you a better grasp of how a ray goes through a scene to get the final image.

Usings Verold's 3ds max exporter - animation not working

If I export the max project as an .FBX it will work (HERE - the pedal depresses ok), but then I loose turbo smooth (wheel alloys) and line objects (cables) disappear (why is that btw?). Using the Verold exporter, the animation asset appears in the 'key frame animation' tab in the 'choose asset' section, but when applied it never actually animates. And yet, if I animate a simple box next to the pedal, that works (HERE).
Partial solution: It seems the problem for Verold was that the pedal's parts were grouped, it works if they are attached instead. –
To clarify, it sounds like there are three issues:
Your TurboSmooth modifier is not being applied when exporting to FBX
Your line objects are not rendering
Your animation is not playing correctly, when exported by Verold Publisher
To apply TurboSmooth when exporting to FBX, please see the following:
http://answers.unity3d.com/questions/23085/3ds-max-fbx-turbosmooth-export.html
Note that line objects are not currently supported in Verold Studio, so they will not render.
Regarding animation, when I look at your first example, I see an animated pedal. In your second example, I see an animated box. Should there be additional animated objects in either example?

Resizing CATiledLayer's Using Scale Transforms vs. Bounds Manipulation

I've got my layer hosted workspace working so that using CATiledLayers for hundreds of images works nicely when the workspace is zoomed out substantially. All the images use lower resolution representations, and my application is much more responsive when panning and zooming large numbers of images.
However, within my application I also provide the user the ability to resize layers with a resize handle. Before I converted image layers to use CATiledLayers I was doing layer resizes by manipulating the bounds of the image layer according to the resize delta (mouse drag), and it worked well. But now with CATiledLayers in place, this is causing CATiledLayers to get confused when I mix resizing of layers through bounds manipulation and zooming/unzooming the workspace through scale transforms.
Specifically, if I resize a CATiledLayer to half the width/height size (1/4 the area), the image inside it will suddenly scale to a further 1/2 the resized frame leaving 3/4 of the frame empty. This seems to be exactly when the inner CATiledLayer logic gets invoked to provide a lower resolution image representation. It works fine if I don't touch the resize handler and just zoom/unzoom the workspace.
Is there a way to make zooming/resizing play nice together with CATiledLayers, or am I going to have to convert my layer resize logic to use scale transforms instead of bounds manipulations?
I ended up solving this by converting my layer resize logic to use scale transforms by overriding the setBounds: method for my custom image layer class to scale it's containing CATiledLayer, and repositioning accordingly. Also it is important to make sure the CATiledLayer's autoresizingMask is set to kCALayerNotSizable since we are handling resizes manually in setBounds:.
Note: be sure to call the superclass's implementation of setBounds:.