Close Menu

    Subscribe to Updates

    Get the latest news from tastytech.

    What's Hot

    La Liga Soccer: Stream Real Madrid vs. Atlético Madrid Live From Anywhere

    March 22, 2026

    How Resident Evil Shifted Perspectives And Framed Fear Over 30 Years

    March 22, 2026

    The Meffs- Business

    March 22, 2026
    Facebook X (Twitter) Instagram
    Facebook X (Twitter) Instagram
    tastytech.intastytech.in
    Subscribe
    • AI News & Trends
    • Tech News
    • AI Tools
    • Business & Startups
    • Guides & Tutorials
    • Tech Reviews
    • Automobiles
    • Gaming
    • movies
    tastytech.intastytech.in
    Home»Business & Startups»I Asked ChatGPT, Claude and DeepSeek to Build Tetris
    I Asked ChatGPT, Claude and DeepSeek to Build Tetris
    Business & Startups

    I Asked ChatGPT, Claude and DeepSeek to Build Tetris

    gvfx00@gmail.comBy gvfx00@gmail.comJanuary 6, 2026No Comments8 Mins Read
    Share
    Facebook Twitter LinkedIn Pinterest Email


    I Asked ChatGPT, Claude and DeepSeek to Build Tetris
    Image by Author

     

    Table of Contents

    Toggle
    • # Introduction
    • # The Prompt
    • # The Results
        • // 1. Claude Opus 4.5
        • // 2. GPT-5.2 Pro
        • // 3. DeepSeek V3.2
    • # Summary: GPT-5.2 vs Opus 4.5 vs DeepSeek 3.2
        • // Cost Breakdown
    • # Takeaways
      • Related posts:
    • 15 Steps to Ensure Your Company’s Compliance
    • How Do You Estimate The Time And Cost Of A Machine Learning Project?
    • 5 Open Source Image Editing AI Models

    # Introduction

     
    It seems like almost every week, a new model claims to be state-of-the-art, beating existing AI models on all benchmarks.

    I get free access to the latest AI models at my full-time job within weeks of release. I typically don’t pay much attention to the hype and just use whichever model is auto-selected by the system.

    However, I know developers and friends who want to build software with AI that can be shipped to production. Since these initiatives are self-funded, their challenge lies in finding the best model to do the job. They want to balance cost with reliability.

    Due to this, after the release of GPT-5.2, I decided to run a practical test to understand whether this model was worth the hype, and if it really was better than the competition.

    Specifically, I chose to test flagship models from each provider: Claude Opus 4.5 (Anthropic’s most capable model), GPT-5.2 Pro (OpenAI’s latest extended reasoning model), and DeepSeek V3.2 (one of the latest open-source alternatives).

    To put these models to the test, I chose to get them to build a playable Tetris game with a single prompt.

    These were the metrics I used to evaluate the success of each model:

     

    Criteria Description
    First Attempt Success With just one prompt, did the model deliver working code? Multiple debugging iterations leads to higher cost over time, which is why this metric was chosen.
    Feature Completeness Were all the features mentioned in the prompt built by the model, or was anything missed out?
    Playability Beyond the technical implementation, was the game actually smooth to play? Or were there issues that created friction in the user experience?
    Cost-effectiveness How much did it cost to get production-ready code?

     

    # The Prompt

     
    Here is the prompt I entered into each AI model:

    Build a fully functional Tetris game as a single HTML file that I can open directly in my browser.

    Requirements:

    GAME MECHANICS:
    – All 7 Tetris piece types
    – Smooth piece rotation with wall kick collision detection
    – Pieces should fall automatically, increase the speed gradually as the user’s score increases
    – Line clearing with visual animation
    – “Next piece” preview box
    – Game over detection when pieces reach the top

    CONTROLS:
    – Arrow keys: Left/Right to move, Down to drop faster, Up to rotate
    – Touch controls for mobile: Swipe left/right to move, swipe down to drop, tap to rotate
    – Spacebar to pause/unpause
    – Enter key to restart after game over

    VISUAL DESIGN:
    – Gradient colors for each piece type
    – Smooth animations when pieces move and lines clear
    – Clean UI with rounded corners
    – Update scores in real time
    – Level indicator
    – Game over screen with final score and restart button

    GAMEPLAY EXPERIENCE AND POLISH:
    – Smooth 60fps gameplay
    – Particle effects when lines are cleared (optional but impressive)
    – Increase the score based on number of lines cleared simultaneously
    – Grid background
    – Responsive design

    Make it visually polished and feel satisfying to play. The code should be clean and well-organized.

     

     

    # The Results

     

    // 1. Claude Opus 4.5

    The Opus 4.5 model built exactly what I asked for.

    The UI was clean and instructions were displayed clearly on the screen. All the controls were responsive and the game was fun to play.

    The gameplay was so smooth that I actually ended up playing for quite some time and got sidetracked from testing the other models.

    Also, Opus 4.5 took less than 2 minutes to provide me with this working game, leaving me impressed on the first try.

     

    Tetris Gameplay Screen by ClaudeTetris Gameplay Screen by Claude
    Tetris game built by Opus 4.5

     

    // 2. GPT-5.2 Pro

    GPT-5.2 Pro is OpenAI’s latest model with extended reasoning. For context, GPT-5.2 has three tiers: Instant, Thinking, and Pro. At the point of writing this article, GPT-5.2 Pro is their most intelligent model, providing extended thinking and reasoning capabilities.

    It is also 4x more expensive than Opus 4.5.

    There was a lot of hype around this model, leading me to go in with high expectations.

    Unfortunately, I was underwhelmed by the game this model produced.

    At the first try, GPT-5.2 Pro produced a Tetris game with a layout bug. The bottom rows of the game were outside of the viewport, and I couldn’t see where the pieces were landing.

    This made the game unplayable, as shown in the screenshot below:

     

    Tetris game built by GPT-5.2Tetris game built by GPT-5.2
    Tetris game built by GPT-5.2

     

    I was especially surprised by this bug since it took around 6 minutes for the model to produce this code.

    I decided to try again with this follow-up prompt to fix the viewport problem:

    The game works, but there’s a bug. The bottom rows of the Tetris board are cut off at the bottom of the screen. I can’t see the pieces when they land and the canvas extends beyond the visible viewport.

    Please fix this by:
    1. Making sure the entire game board fits in the viewport
    2. Adding proper centering so the full board is visible

    The game should fit on the screen with all rows visible.

     

    After the follow-up prompt, the GPT-5.2 Pro model produced a functional game, as seen in the below screenshot:

     

    Tetris Second Try by GPT-5.2Tetris Second Try by GPT-5.2
    Tetris second try by GPT-5.2

     

    However, the game play wasn’t as smooth as the one produced by the Opus 4.5 model.

    When I pressed the “down” arrow for the piece to drop, the next piece would sometimes plummet instantly at a high speed, not giving me enough time to think about how to position it.

    The game ended up being playable only if I let each piece fall on its own, which wasn’t the best experience.

    (Note: I tried the GPT-5.2 Standard model too, which produced similar buggy code on the first try.)

     

    // 3. DeepSeek V3.2

    DeepSeek’s first attempt at building this game had two issues:

    • Pieces started disappearing when they hit the bottom of the screen.
    • The “down” arrow that’s used to drop the pieces faster ended up scrolling the entire webpage rather than just moving the game pieces.

     

    Tetris game built by DeepSeek V3.2Tetris game built by DeepSeek V3.2
    Tetris game built by DeepSeek V3.2

     

    I re-prompted the model to fix this issue, and the gameplay controls ended up working correctly.

    However, some pieces still disappeared before they landed. This made the game completely unplayable even after the second iteration.

    I’m sure that this issue can be fixed with 2–3 more prompts, and given DeepSeek’s low pricing, you could afford 10+ debugging rounds and still spend less than one successful Opus 4.5 attempt.

     

    # Summary: GPT-5.2 vs Opus 4.5 vs DeepSeek 3.2

     

    // Cost Breakdown

    Here is a cost comparison between the three models:
     

    Model Input (per 1M tokens) Output (per 1M tokens)
    DeepSeek V3.2 $0.27 $1.10
    GPT-5.2 $1.75 $14.00
    Claude Opus 4.5 $5.00 $25.00
    GPT-5.2 Pro $21.00 $84.00

     

    DeepSeek V3.2 is the cheapest alternative, and you can also download the model’s weights for free and run it on your own infrastructure.

    GPT-5.2 is almost 7x more expensive than DeepSeek V3.2, followed by Opus 4.5 and GPT-5.2 Pro.

    For this specific task (building a Tetris game), we consumed approximately 1,000 input tokens and 3,500 output tokens.

    For each additional iteration, we will estimate an extra 1,500 tokens per additional round. Here is the total cost incurred per model:

     

    Model Total Cost Result
    DeepSeek V3.2 ~$0.005 Game isn’t playable
    GPT-5.2 ~$0.07 Playable, but poor user experience
    Claude Opus 4.5 ~$0.09 Playable and good user experience
    GPT-5.2 Pro ~$0.41 Playable, but poor user experience

     

    # Takeaways

     
    Based on my experience building this game, I would stick to the Opus 4.5 model for day to day coding tasks.

    Although GPT-5.2 is cheaper than Opus 4.5, I personally would not use it to code, since the iterations required to yield the same result would likely lead to the same amount of money spent.

    DeepSeek V3.2, however, is far more affordable than the other models on this list.

    If you’re a developer on a budget and have time to spare on debugging, you will still end up saving money even if it takes you over 10 tries to get working code.

    I was surprised at GPT 5.2 Pro’s inability to produce a working game on the first try, as it took around 6 minutes to think before coming up with flawed code. After all, this is OpenAI’s flagship model, and Tetris should be a relatively simple task.

    However, GPT-5.2 Pro’s strengths lie in math and scientific research, and it is especially designed for problems that don’t rely on pattern recognition from training data. Perhaps this model is over-engineered for simple day-to-day coding tasks, and should instead be used when building something that is complex and requires novel architecture.

    The practical takeaway from this experiment:

    • Opus 4.5 performs best at day-to-day coding tasks.
    • DeepSeek V3.2 is a budget alternative that delivers reasonable output, although it requires some debugging effort to reach your desired outcome.
    • GPT-5.2 (Standard) didn’t perform as well as Opus 4.5, whereas GPT-5.2 (Pro) is probably better suited for complex reasoning than quick coding tasks like this one.

    Feel free to replicate this test with the prompt I’ve shared above, and happy coding!
    &nbsp
    &nbsp

    Natassha Selvaraj is a self-taught data scientist with a passion for writing. Natassha writes on everything data science-related, a true master of all data topics. You can connect with her on LinkedIn or check out her YouTube channel.

    Related posts:

    What is artificial intelligence? - The pragmatic definition — Dan Rose AI

    Top AI Blogs and Websites To Follow in 2024

    The Algorithmic X-Men - KDnuggets

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticlePeacock will be the first streaming service with Dolby Vision 2 HDR and next-gen Atmos audio encoding
    Next Article Americans evenly split on Maduro’s abduction, poll shows | Donald Trump News
    gvfx00@gmail.com
    • Website

    Related Posts

    Business & Startups

    5 Useful Python Scripts for Synthetic Data Generation

    March 21, 2026
    Business & Startups

    The Better Way For Document Chatbots?

    March 21, 2026
    Business & Startups

    5 Powerful Python Decorators for Robust AI Agents

    March 21, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    BMW Will Put eFuel In Cars Made In Germany From 2028

    October 14, 202511 Views

    Best Sonic Lego Deals – Dr. Eggman’s Drillster Gets Big Price Cut

    December 16, 20259 Views

    What is Fine-Tuning? Your Ultimate Guide to Tailoring AI Models in 2025

    October 14, 20259 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram

    Subscribe to Updates

    Get the latest tech news from tastytech.

    About Us
    About Us

    TastyTech.in brings you the latest AI, tech news, cybersecurity tips, and gadget insights all in one place. Stay informed, stay secure, and stay ahead with us!

    Most Popular

    BMW Will Put eFuel In Cars Made In Germany From 2028

    October 14, 202511 Views

    Best Sonic Lego Deals – Dr. Eggman’s Drillster Gets Big Price Cut

    December 16, 20259 Views

    What is Fine-Tuning? Your Ultimate Guide to Tailoring AI Models in 2025

    October 14, 20259 Views

    Subscribe to Updates

    Get the latest news from tastytech.

    Facebook X (Twitter) Instagram Pinterest
    • Homepage
    • About Us
    • Contact Us
    • Privacy Policy
    © 2026 TastyTech. Designed by TastyTech.

    Type above and press Enter to search. Press Esc to cancel.

    Ad Blocker Enabled!
    Ad Blocker Enabled!
    Our website is made possible by displaying online advertisements to our visitors. Please support us by disabling your Ad Blocker.