Subscribe to PixelSham.com RSS for free

3Dprinting (179) A.I. (897) animation (353) blender (217) colour (240) commercials (53) composition (154) cool (368) design (655) Featured (91) hardware (316) IOS (109) jokes (140) lighting (298) modeling (156) music (189) photogrammetry (197) photography (757) production (1308) python (101) quotes (498) reference (317) software (1379) trailers (308) ves (571) VR (221)

POPULAR SEARCHES unreal | pipeline | virtual production | free | learn | photoshop | 360 | macro | google | nvidia | resolution | open source | hdri | real-time | photography basics | nuke

  • Explore Posts
  • Job Postings
  • ReelMatters.com
  • About and Contact
    • About And Contact
    • Portfolio
    • Privacy Policy
    • RSS feed page

BREAKING NEWS

LATEST POSTS

  • Decart AI Mirage – The first ever World Transformation Model – turning any video, game, or camera feed into a new digital world, in real time

    pIXELsHAM.com
    Jul 25, 2025
    Featured

    https://mirage.decart.ai/

    https://about.decart.ai/publications/mirage

    Views : 13
  • Mike Wong – AtoMeow – A Blue noise image stippling in Processing

    pIXELsHAM.com
    Jul 24, 2025
    design, software

    https://github.com/mwkm/atoMeow

    https://www.shadertoy.com/view/7s3XzX

    This demo is created for coders who are familiar with this awesome creative coding platform. You may quickly modify the code to work for video or to stipple your own Procssing drawings by turning them into PImage and run the simulation. This demo code also serves as a reference implementation of my article Blue noise sampling using an N-body simulation-based method. If you are interested in 2.5D, you may mod the code to achieve what I discussed in this artist friendly article.

    Convert your video to a dotted noise.

    Views : 29
  • Aitor Echeveste – Free CG and Comp Projection Shot, Download the Assets & Follow the Workflow

    pIXELsHAM.com
    Jul 24, 2025
    production, reference

    What’s Included:

    • Cleaned and extended base plates
    • Full Maya and Nuke 3D projection layouts
    • Bullet and environment CG renders with AOVs (RGB, normals, position, ID, etc.)
    • Explosion FX in slow motion
    • 3D scene geometry for projection
    • Camera + lensing setup
    • Light groups and passes for look development

    https://aitorecheveste.gumroad.com/l/flbeqr

    Views : 17
  • Tauseef Fayyaz About readable code – Clean Code Practices

    pIXELsHAM.com
    Jul 24, 2025
    python

    𝗛𝗲𝗿𝗲’𝘀 𝘄𝗵𝗮𝘁 𝘁𝗼 𝗺𝗮𝘀𝘁𝗲𝗿 𝗶𝗻 𝗖𝗹𝗲𝗮𝗻 𝗖𝗼𝗱𝗲 𝗣𝗿𝗮𝗰𝘁𝗶𝗰𝗲𝘀:

    🔹 Code Readability & Simplicity – Use meaningful names, write short functions, follow SRP, flatten logic, and remove dead code.
    → Clarity is a feature.

    🔹 Function & Class Design – Limit parameters, favor pure functions, small classes, and composition over inheritance.
    → Structure drives scalability.

    🔹 Testing & Maintainability – Write readable unit tests, avoid over-mocking, test edge cases, and refactor with confidence.
    → Test what matters.

    🔹 Code Structure & Architecture – Organize by features, minimize global state, avoid god objects, and abstract smartly.
    → Architecture isn’t just backend.

    🔹 Refactoring & Iteration – Apply the Boy Scout Rule, DRY, KISS, and YAGNI principles regularly.
    → Refactor like it’s part of development.

    🔹 Robustness & Safety – Validate early, handle errors gracefully, avoid magic numbers, and favor immutability.
    → Safe code is future-proof.

    🔹 Documentation & Comments – Let your code explain itself. Comment why, not what, and document at the source.
    → Good docs reduce team friction.

    🔹 Tooling & Automation – Use linters, formatters, static analysis, and CI reviews to automate code quality.
    → Let tools guard your gates.

    🔹 Final Review Practices – Review, refactor nearby code, and avoid cleverness in the name of brevity.
    → Readable code is better than smart code.

    Tauseef Fayyaz About readable code – Clean Code Practices_compressedDownload
    Views : 11
  • Mark Theriault “Steamboat Willie” – AI Re-Imagining of a 1928 Classic in 4k

    pIXELsHAM.com
    Jul 24, 2025
    A.I., trailers

    I ran Steamboat Willie (now public domain) through Flux Kontext to reimagine it as a 3D-style animated piece. Instead of going the polished route with something like W.A.N. 2.1 for full image-to-video generation, I leaned into the raw, handmade vibe that comes from converting each frame individually. It gave it a kind of stop-motion texture, imperfect, a bit wobbly, but full of character.

    Views : 23
  • Glen Keane- Tarzan Construction Sheets + Chalk Talks

    pIXELsHAM.com
    Jul 24, 2025
    animation, design
    Glen Keane- Tarzan Construction Sheets + Chalk TalksDownload
    Views : 9
  • Microsoft DAViD – Data-efficient and Accurate Vision Models from Synthetic Data

    pIXELsHAM.com
    Jul 22, 2025
    A.I., software

    Our human-centric dense prediction model delivers high-quality, detailed (depth) results while achieving remarkable efficiency, running orders of magnitude faster than competing methods, with inference speeds as low as 21 milliseconds per frame (the large multi-task model on an NVIDIA A100). It reliably captures a wide range of human characteristics under diverse lighting conditions, preserving fine-grained details such as hair strands and subtle facial features. This demonstrates the model’s robustness and accuracy in complex, real-world scenarios.

    https://microsoft.github.io/DAViD

    The state of the art in human-centric computer vision achieves high accuracy and robustness across a diverse range of tasks. The most effective models in this domain have billions of parameters, thus requiring extremely large datasets, expensive training regimes, and compute-intensive inference. In this paper, we demonstrate that it is possible to train models on much smaller but high-fidelity synthetic datasets, with no loss in accuracy and higher efficiency. Using synthetic training data provides us with excellent levels of detail and perfect labels, while providing strong guarantees for data provenance, usage rights, and user consent. Procedural data synthesis also provides us with explicit control on data diversity, that we can use to address unfairness in the models we train. Extensive quantitative assessment on real input images demonstrates accuracy of our models on three dense prediction tasks: depth estimation, surface normal estimation, and soft foreground segmentation. Our models require only a fraction of the cost of training and inference when compared with foundational models of similar accuracy.

    Views : 15
  • VEO3 – Ads’ prompt examples

    pIXELsHAM.com
    Jul 22, 2025
    A.I., commercials

    https://www.linkedin.com/posts/leokadieff_ai-generativeai-filmmaking-activity-7353474389029330950-luom

    Prompts and more examples under the post

    (more…)
    Views : 51
  • Stability Matrix for ComfyUI and similar genAI apps

    pIXELsHAM.com
    Jul 22, 2025
    A.I., software

    https://github.com/LykosAI/StabilityMatrix

    Views : 10
  • Embedding frame ranges into Quicktime movies with FFmpeg

    pIXELsHAM.com
    Jul 22, 2025
    Featured, software

    QuickTime (.mov) files are fundamentally time-based, not frame-based, and so don’t have a built-in, uniform “first frame/last frame” field you can set as numeric frame IDs. Instead, tools like Shotgun Create rely on the timecode track and the movie’s duration to infer frame numbers. If you want Shotgun to pick up a non-default frame range (e.g. start at 1001, end at 1064), you must bake in an SMPTE timecode that corresponds to your desired start frame, and ensure the movie’s duration matches your clip length.

    How Shotgun Reads Frame Ranges

    • Default start frame is 1. If no timecode metadata is present, Shotgun assumes the movie begins at frame 1.
    • Timecode ⇒ frame number. Shotgun Create “honors the timecodes of media sources,” mapping the embedded TC to frame IDs. For example, a 24 fps QuickTime tagged with a start timecode of 00:00:41:17 will be interpreted as beginning on frame 1001 (1001 ÷ 24 fps ≈ 41.71 s).

    Embedding a Start Timecode

    QuickTime uses a tmcd (timecode) track. You can bake in an SMPTE track via FFmpeg’s -timecode flag or via Compressor/encoder settings:

    1. Compute your start TC.
      • Desired start frame = 1001
      • Frame 1001 at 24 fps ⇒ 1001 ÷ 24 ≈ 41.708 s ⇒ TC 00:00:41:17
    2. FFmpeg example:
    ffmpeg -i input.mov \
      -c copy \
      -timecode 00:00:41:17 \
      output.mov
    

    This adds a timecode track beginning at 00:00:41:17, which Shotgun maps to frame 1001.

    Ensuring the Correct End Frame

    Shotgun infers the last frame from the movie’s duration. To end on frame 1064:

    • Frame count = 1064 – 1001 + 1 = 64 frames
    • Duration = 64 ÷ 24 fps ≈ 2.667 s

    FFmpeg trim example:

    ffmpeg -i input.mov \
      -c copy \
      -timecode 00:00:41:17 \
      -t 00:00:02.667 \
      output_trimmed.mov
    

    This results in a 64-frame clip (1001→1064) at 24 fps.

    Views : 6
  • Aider.chat – A free, open-source AI pair-programming CLI tool

    pIXELsHAM.com
    Jul 19, 2025
    A.I., software

    https://aider.chat/

    Aider enables developers to interactively generate, modify, and test code by leveraging both cloud-hosted and local LLMs directly from the terminal or within an IDE. Key capabilities include comprehensive codebase mapping, support for over 100 programming languages, automated git commit messages, voice-to-code interactions, and built-in linting and testing workflows. Installation is straightforward via pip or uv, and while the tool itself has no licensing cost, actual usage costs stem from the underlying LLM APIs, which are billed separately by providers like OpenAI or Anthropic.

    Key Features

    • Cloud & Local LLM Support
      Connect to most major LLM providers out of the box, or run models locally for privacy and cost control aider.chat.
    • Codebase Mapping
      Automatically indexes all project files so that even large repositories can be edited contextually aider.chat.
    • 100+ Language Support
      Works with Python, JavaScript, Rust, Ruby, Go, C++, PHP, HTML, CSS, and dozens more aider.chat.
    • Git Integration
      Generates sensible commit messages and automates diffs/undo operations through familiar git tooling aider.chat.
    • Voice-to-Code
      Speak commands to Aider to request features, tests, or fixes without typing aider.chat.
    • Images & Web Pages
      Attach screenshots, diagrams, or documentation URLs to provide visual context for edits aider.chat.
    • Linting & Testing
      Runs lint and test suites automatically after each change, and can fix issues it detects
    (more…)
    Views : 20
  • DJI adds Gaussian Splatting support to DJI Terra

    pIXELsHAM.com
    Jul 18, 2025
    hardware, photogrammetry, software

    https://enterprise.dji.com/dji-terra

    Views : 9
  • Netflix starts using GenAI in its shows and films

    pIXELsHAM.com
    Jul 18, 2025
    A.I., ves

    https://techcrunch.com/2025/07/18/netflix-starts-using-genai-in-its-shows-and-films/

    Views : 7
  • SourceTree vs Github Desktop – Which one to use

    pIXELsHAM.com
    Jul 17, 2025
    Featured, software

    Sourcetree and GitHub Desktop are both free, GUI-based Git clients aimed at simplifying version control for developers. While they share the same core purpose—making Git more accessible—they differ in features, UI design, integration options, and target audiences.


    Installation & Setup

    • Sourcetree
      • Download: https://www.sourcetreeapp.com/
      • Supported OS: Windows 10+, macOS 10.13+
      • Prerequisites: Comes bundled with its own Git, or can be pointed to a system Git install.
      • Initial Setup: Wizard guides SSH key generation, authentication with Bitbucket/GitHub/GitLab.
    • GitHub Desktop
      • Download: https://desktop.github.com/
      • Supported OS: Windows 10+, macOS 10.15+
      • Prerequisites: Bundled Git; seamless login with GitHub.com or GitHub Enterprise.
      • Initial Setup: One-click sign-in with GitHub; auto-syncs repositories from your GitHub account.

    Feature Comparison

    FeatureSourcetreeGitHub Desktop
    Branch VisualizationDetailed graph view with drag-and-drop for rebasing/mergingLinear graph, simpler but less configurable
    Staging & CommitFile-by-file staging, inline diff viewAll-or-nothing staging, side-by-side diff
    Interactive RebaseFull support via UIBasic support via command line only
    Conflict ResolutionBuilt-in merge tool integration (DiffMerge, Beyond Compare)Contextual conflict editor with choice panels
    Submodule ManagementNative submodule supportLimited; requires CLI
    Custom Actions / HooksDefine custom actions (e.g., launch scripts)No UI for custom Git hooks
    Git Flow / Hg FlowBuilt-in supportNone
    PerformanceCan lag on very large reposGenerally snappier on medium-sized repos
    Memory FootprintHigher RAM usageLightweight
    Platform IntegrationAtlassian Bitbucket, JiraDeep GitHub.com / Enterprise integration
    Learning CurveSteeper for beginnersBeginner-friendly
    (more…)
    Views : 114
  • Jeff Leu – The Cinematography of Roger Deakins – How His Visual Storytelling Reflects His Philosophies

    pIXELsHAM.com
    Jul 16, 2025
    composition, lighting

    https://eloncdn.blob.core.windows.net/eu3/sites/153/2020/06/11-Leu.pdf

    The Cinematography of Roger DeakinsDownload
    Views : 8
Previous Page
1 … 5 6 7 8 9 … 433
Next Page

FEATURED POSTS

  • Wolfcrow – Top 100 Films for Gorgeous Cinematography

    pIXELsHAM.com
    May 10, 2022
    composition, lighting, photography

    Views : 671
  • NVidia Neural Physics Enables Realistic Real Time Hair Simulations

    pIXELsHAM.com
    May 2, 2023
    A.I., software

     

    https://blogs.nvidia.com/blog/2023/05/02/graphics-research-advances-generative-ai-next-frontier

    Views : 489
  • Light Controls De-Mystified

    pIXELsHAM.com
    May 3, 2014
    lighting, photography

    http://www.lowel.com/edu/light_controls/index.html

    Views : 1,057
  • Sensitivity of human eye

    pIXELsHAM.com
    Mar 10, 2016
    colour, Featured, photography, reference

    http://www.wikilectures.eu/index.php/Spectral_sensitivity_of_the_human_eye

    http://www.normankoren.com/Human_spectral_sensitivity_small.jpg

    Spectral sensitivity of eye is influenced by light intensity. And the light intensity determines the level of activity of cones cell and rod cell. This is the main characteristic of human vision. Sensitivity to individual colors, in other words, wavelengths of the light spectrum, is explained by the RGB (red-green-blue) theory. This theory assumed that there are three kinds of cones. It’s selectively sensitive to red (700-630 nm), green (560-500 nm), and blue (490-450 nm) light. And their mutual interaction allow to perceive all colors of the spectrum.

    http://weeklysciencequiz.blogspot.com/2013/01/violet-skies-are-for-birds.html

     

    (more…)
    Views : 8,113
  • Yann Lecun: Meta AI, Open Source, Limits of LLMs, AGI & the Future of AI | Lex Fridman Podcast #416

    pIXELsHAM.com
    Mar 14, 2024
    A.I., Featured, quotes

    Views : 138
  • Akiyoshi Kitaoka – Color perception. Is this Coke can really red?

    pIXELsHAM.com
    Mar 30, 2024
    colour

     

    Views : 239
  • Japanese Designer Tomoo Yamaji Offers 3D Printed Transformer Kit, Stingray, Through Shapeways

    pIXELsHAM.com
    Jan 3, 2017
    3Dprinting, design

    https://3dprint.com/55799/transformer-kit-shapeways/

    http://www.shapeways.com/product/5YHJL6XSZ/t060101-stingray?li=shareProduct

    Views : 1,648
  • How to Convert Temperature (K) to RGB: Algorithm and Sample Code

    pIXELsHAM.com
    Aug 30, 2021
    colour, lighting, production, software

    tannerhelland.com/2012/09/18/convert-temperature-rgb-algorithm-code.html

    www.vendian.org/mncharity/dir3/blackbody/UnstableURLs/bbr_color.html

    Views : 928
Views : 10,089

RSS feed page

Search


Categories


Archive


Disclaimer


Links and images on this website may be protected by the respective owners’ copyright. All data submitted by users through this site shall be treated as freely available to share.