COMPOSITION
DESIGN
- 
VQGAN + CLIP AI made Music Video for the song Canvas by ResonateRead more: VQGAN + CLIP AI made Music Video for the song Canvas by Resonate” In this video, I utilized artificial intelligence to generate an animated music video for the song Canvas by Resonate. This tool allows anyone to generate beautiful images using only text as the input. My question was, what if I used song lyrics as input to the AI, can I make perfect music synchronized videos automatically with the push of a button? Let me know how you think the AI did in this visual interpretation of the song. After getting caught up in the excitement around DALL·E2 (latest and greatest AI system, it’s INSANE), I searched for any way I could use similar image generation for music synchronization. Since DALL·E2 is not available to the public yet, my search led me to VQGAN + CLIP (Vector Quantized Generative Adversarial Network and Contrastive Language–Image Pre-training), before settling more specifically on Disco Diffusion V5.2 Turbo. If you don’t know what any of these words or acronyms mean, don’t worry, I was just as confused when I first started learning about this technology. I believe we’re reaching a turning point where entire industries are about to shift in reaction to this new process (which is essentially magic!). DoodleChaos” 
COLOR
- 
Björn Ottosson – OKlch color spaceRead more: Björn Ottosson – OKlch color spaceBjörn Ottosson proposed OKlch in 2020 to create a color space that can closely mimic how color is perceived by the human eye, predicting perceived lightness, chroma, and hue. The OK in OKLCH stands for Optimal Color. - L: Lightness (the perceived brightness of the color)
- C: Chroma (the intensity or saturation of the color)
- H: Hue (the actual color, such as red, blue, green, etc.)
  Also read: 
- 
Photography Basics : Spectral Sensitivity Estimation Without a CameraRead more: Photography Basics : Spectral Sensitivity Estimation Without a Camerahttps://color-lab-eilat.github.io/Spectral-sensitivity-estimation-web/ A number of problems in computer vision and related fields would be mitigated if camera spectral sensitivities were known. As consumer cameras are not designed for high-precision visual tasks, manufacturers do not disclose spectral sensitivities. Their estimation requires a costly optical setup, which triggered researchers to come up with numerous indirect methods that aim to lower cost and complexity by using color targets. However, the use of color targets gives rise to new complications that make the estimation more difficult, and consequently, there currently exists no simple, low-cost, robust go-to method for spectral sensitivity estimation that non-specialized research labs can adopt. Furthermore, even if not limited by hardware or cost, researchers frequently work with imagery from multiple cameras that they do not have in their possession. To provide a practical solution to this problem, we propose a framework for spectral sensitivity estimation that not only does not require any hardware (including a color target), but also does not require physical access to the camera itself. Similar to other work, we formulate an optimization problem that minimizes a two-term objective function: a camera-specific term from a system of equations, and a universal term that bounds the solution space. Different than other work, we utilize publicly available high-quality calibration data to construct both terms. We use the colorimetric mapping matrices provided by the Adobe DNG Converter to formulate the camera-specific system of equations, and constrain the solutions using an autoencoder trained on a database of ground-truth curves. On average, we achieve reconstruction errors as low as those that can arise due to manufacturing imperfections between two copies of the same camera. We provide predicted sensitivities for more than 1,000 cameras that the Adobe DNG Converter currently supports, and discuss which tasks can become trivial when camera responses are available.  
- 
Yasuharu YOSHIZAWA – Comparison of sRGB vs ACREScg in NukeRead more: Yasuharu YOSHIZAWA – Comparison of sRGB vs ACREScg in NukeAnswering the question that is often asked, “Do I need to use ACEScg to display an sRGB monitor in the end?” (Demonstration shown at an in-house seminar) 
 Comparison of scanlineRender output with extreme color lights on color charts with sRGB/ACREScg in color – OCIO -working space in NukeDownload the Nuke script: 
- 
VES Cinematic Color – Motion-Picture Color ManagementRead more: VES Cinematic Color – Motion-Picture Color ManagementThis paper presents an introduction to the color pipelines behind modern feature-film visual-effects and animation. Authored by Jeremy Selan, and reviewed by the members of the VES Technology Committee including Rob Bredow, Dan Candela, Nick Cannon, Paul Debevec, Ray Feeney, Andy Hendrickson, Gautham Krishnamurti, Sam Richards, Jordan Soles, and Sebastian Sylwan. 
- 
Tobia Montanari – Memory Colors: an essential tool for ColoristsRead more: Tobia Montanari – Memory Colors: an essential tool for Coloristshttps://www.tobiamontanari.com/memory-colors-an-essential-tool-for-colorists/ “Memory colors are colors that are universally associated with specific objects, elements or scenes in our environment. They are the colors that we expect to see in specific situations: these colors are based on our expectation of how certain objects should look based on our past experiences and memories. For instance, we associate specific hues, saturation and brightness values with human skintones and a slight variation can significantly affect the way we perceive a scene. Similarly, we expect blue skies to have a particular hue, green trees to be a specific shade and so on. Memory colors live inside of our brains and we often impose them onto what we see. By considering them during the grading process, the resulting image will be more visually appealing and won’t distract the viewer from the intended message of the story. Even a slight deviation from memory colors in a movie can create a sense of discordance, ultimately detracting from the viewer’s experience.” 
LIGHTING
- 
Photography basics: Solid Angle measuresRead more: Photography basics: Solid Angle measureshttp://www.calculator.org/property.aspx?name=solid+angle A measure of how large the object appears to an observer looking from that point. Thus. A measure for objects in the sky. Useful to retuen the size of the sun and moon… and in perspective, how much of their contribution to lighting. Solid angle can be represented in ‘angular diameter’ as well. http://en.wikipedia.org/wiki/Solid_angle http://www.mathsisfun.com/geometry/steradian.html A solid angle is expressed in a dimensionless unit called a steradian (symbol: sr). By default in terms of the total celestial sphere and before atmospheric’s scattering, the Sun and the Moon subtend fractional areas of 0.000546% (Sun) and 0.000531% (Moon). http://en.wikipedia.org/wiki/Solid_angle#Sun_and_Moon On earth the sun is likely closer to 0.00011 solid angle after athmospheric scattering. The sun as perceived from earth has a diameter of 0.53 degrees. This is about 0.000064 solid angle. http://www.numericana.com/answer/angles.htm The mean angular diameter of the full moon is 2q = 0.52° (it varies with time around that average, by about 0.009°). This translates into a solid angle of 0.0000647 sr, which means that the whole night sky covers a solid angle roughly one hundred thousand times greater than the full moon. More info http://lcogt.net/spacebook/using-angles-describe-positions-and-apparent-sizes-objects http://amazing-space.stsci.edu/glossary/def.php.s=topic_astronomy Angular Size The apparent size of an object as seen by an observer; expressed in units of degrees (of arc), arc minutes, or arc seconds. The moon, as viewed from the Earth, has an angular diameter of one-half a degree. The angle covered by the diameter of the full moon is about 31 arcmin or 1/2°, so astronomers would say the Moon’s angular diameter is 31 arcmin, or the Moon subtends an angle of 31 arcmin. 
- 
Disney’s Moana Island Scene – Free data setRead more: Disney’s Moana Island Scene – Free data sethttps://www.disneyanimation.com/resources/moana-island-scene/ This data set contains everything necessary to render a version of the Motunui island featured in the 2016 film Moana. 
- 
7 Easy Portrait Lighting SetupsRead more: 7 Easy Portrait Lighting SetupsButterfly Loop Rembrandt Split Rim Broad Short 
- 
About green screensRead more: About green screenshackaday.com/2015/02/07/how-green-screen-worked-before-computers/ www.newtek.com/blog/tips/best-green-screen-materials/ www.chromawall.com/blog//chroma-key-green Chroma Key Green, the color of green screens is also known as Chroma Green and is valued at approximately 354C in the Pantone color matching system (PMS). Chroma Green can be broken down in many different ways. Here is green screen green as other values useful for both physical and digital production: Green Screen as RGB Color Value: 0, 177, 64 
 Green Screen as CMYK Color Value: 81, 0, 92, 0
 Green Screen as Hex Color Value: #00b140
 Green Screen as Websafe Color Value: #009933Chroma Key Green is reasonably close to an 18% gray reflectance. Illuminate your green screen with an uniform source with less than 2/3 EV variation. 
 The level of brightness at any given f-stop should be equivalent to a 90% white card under the same lighting.
COLLECTIONS
| Featured AI
| Design And Composition 
| Explore posts  
POPULAR SEARCHES
unreal | pipeline | virtual production | free | learn | photoshop | 360 | macro | google | nvidia | resolution | open source | hdri | real-time | photography basics | nuke
FEATURED POSTS
- 
Tencent Hunyuan3D 2.1 goes Open Source and adds MV (Multi-view) and MV Mini
- 
FFmpeg – examples and convenience lines
- 
Canva bought Affinity – Now Affinity Photo and Affinity Designer are… GONE?!
- 
Alejandro Villabón and Rafał Kaniewski – Recover Highlights With 8-Bit to High Dynamic Range Half Float Copycat – Nuke
- 
What the Boeing 737 MAX’s crashes can teach us about production business – the effects of commoditisation
- 
Types of AI Explained in a few Minutes – AI Glossary
- 
Photography basics: Shutter angle and shutter speed and motion blur
- 
Daniele Tosti Interview for the magazine InCG, Taiwan, Issue 28, 201609
Social Links
DISCLAIMER – Links and images on this website may be protected by the respective owners’ copyright. All data submitted by users through this site shall be treated as freely available to share.












































