BREAKING NEWS
LATEST POSTS
-
PTGui 13 beta adds control through a Patch Editor
Additions:
- Patch Editor (PTGui Pro)
- DNG output
- Improved RAW / DNG handling
- JPEG 2000 support
- Performance improvements
-
The riddles humans can solve but AI computers cannot
https://www.bbc.com/future/article/20240912-what-riddles-teach-us-about-the-human-mind
“As human beings, it’s very easy for us to have common sense, and apply it at the right time and adapt it to new problems,” says Ilievski, who describes his branch of computer science as “common sense AI”. But right now, AI has a “general lack of grounding in the world”, which makes that kind of basic, flexible reasoning a struggle.
AI excels at pattern recognition, “but it tends to be worse than humans at questions that require more abstract thinking”, says Xaq Pitkow, an associate professor at Carnegie Mellon University in the US, who studies the intersection of AI and neuroscience. In many cases, though, it depends on the problem.
A bizarre truth about AI is we have no idea how it works. The same is true about the brain.
That’s why the best systems may come from a combination of AI and human work; we can play to the machine’s strengths, Ilievski says.
-
AI and the Law – CartoonBrew.com : Lionsgate signs deal with AI company Runway, hoping that AI can eliminate storyboard artists and VFX crews
The goal is to reduce costs by replacing traditional storyboard artists and VFX crews with AI-generated “cinematic video.” Lionsgate hopes to use this technology for both pre- and post-production processes. While the company promotes the cost-saving potential, the creative community has raised concerns, as Runway is currently facing a lawsuit over copyright infringement.
-
Clint Eastwood on the set of his latest movie
At the age of 94, this is what the great Clint Eastwood looks like.
Standing, lucid, brilliant, directing his latest film. Eastwood himself says it: “I don’t let the old man in. I keep myself busy. You have to stay active, alive, happy, strong, capable. I don’t let in the old critic, hostile, envious, gossiping, full of rage and complaints, of lack of courage, which denies to itself that old age can be creative, decisive, full of light and projection. Getting older is not for sissies.”
~Clint Eastwood
-
How to Lead Your Team when the House Is on Fire
https://peterszasz.com/how-to-lead-your-team-when-the-house-is-on-fire/
The three focus areas of an Engineering Manager
- Ensuring delivery that’s aligned with company goals;
- Building and sustaining a high-performing engineering team;
- Supporting the success and personal growth of the individuals on the team.
FEATURED POSTS
-
How does Stable Diffusion work?
https://stable-diffusion-art.com/how-stable-diffusion-work/
Stable Diffusion is a latent diffusion model that generates AI images from text. Instead of operating in the high-dimensional image space, it first compresses the image into the latent space.
Stable Diffusion belongs to a class of deep learning models called diffusion models. They are generative models, meaning they are designed to generate new data similar to what they have seen in training. In the case of Stable Diffusion, the data are images.
Why is it called the diffusion model? Because its math looks very much like diffusion in physics. Let’s go through the idea.
-
Scientists claim to have discovered ‘new colour’ no one has seen before: Olo
https://www.bbc.com/news/articles/clyq0n3em41o
By stimulating specific cells in the retina, the participants claim to have witnessed a blue-green colour that scientists have called “olo”, but some experts have said the existence of a new colour is “open to argument”.
The findings, published in the journal Science Advances on Friday, have been described by the study’s co-author, Prof Ren Ng from the University of California, as “remarkable”.
(A) System inputs. (i) Retina map of 103 cone cells preclassified by spectral type (7). (ii) Target visual percept (here, a video of a child, see movie S1 at 1:04). (iii) Infrared cellular-scale imaging of the retina with 60-frames-per-second rolling shutter. Fixational eye movement is visible over the three frames shown.
(B) System outputs. (iv) Real-time per-cone target activation levels to reproduce the target percept, computed by: extracting eye motion from the input video relative to the retina map; identifying the spectral type of every cone in the field of view; computing the per-cone activation the target percept would have produced. (v) Intensities of visible-wavelength 488-nm laser microdoses at each cone required to achieve its target activation level.
(C) Infrared imaging and visible-wavelength stimulation are physically accomplished in a raster scan across the retinal region using AOSLO. By modulating the visible-wavelength beam’s intensity, the laser microdoses shown in (v) are delivered. Drawing adapted with permission [Harmening and Sincich (54)].
(D) Examples of target percepts with corresponding cone activations and laser microdoses, ranging from colored squares to complex imagery. Teal-striped regions represent the color “olo” of stimulating only M cones.