I recently made a photo post with my impressions of the yellow iPhone 14, which was completely shot on the iPhone 13 Pro.

It’s terrible, but for processing in photoshop Togothat the smartphone issued, killed almost the whole day. And not because my hands are crooked. Aligning all the frames to one color game turned out to be too difficult.

In each new frame, the iPhone painted the tree, hands, and objects in random order. And that’s half the problem.

Below are some sample examples showing where the iPhone sags hard and Apple is just a couple of new tweaks that should lead to a new level of comfort and naturalness.

There is still no manual color fixing in the Camera


Third party apps don’t provide the same convenience

Since 2019, the top iPhone models have been released under the Pro brand. And the stories about the presentation: professional features are now being added that require three percent of buyers from strength, but make everyone want to buy.

He was the first Pro: Review of the iPhone 11 Pro and Pro Max. Green is not for everyone

At the same time, the settings inside the camera applications were frail for the fifth year in a row. Apple, of course, is trying to make them its own way, rather than mindlessly copying them from professional cameras. But because there is still no white balanceI want to score on mobile shooting in the corner.


White balance gradation. The lower the value, the more yellow

white balance responsible for making white look white and image color temperature application of popularity. Photo or video did not go away flashed and blue.

measured in Kelvin. The gradation is:

  • Warm white (2700-3200K)
  • Day white (3500-4500K)
  • White (5000-6000K)
  • Cold white (6000-8000K).
  • On conventional cameras, correct white means that all screens also begin to appear naturally.

    In theory, the iPhone should also work this way, but two problems are revealed at once: you cannot fix the value manually, and even with the same indicator of this parameter, the color will be different.

    The issue with the manual setting is most noticeable when you’re trying to capture a purple sunset and the iPhone is skinning it.

    Two photos without processing, neither of them showed the real shade of the sky

    In the photo above, the truth was somewhere in between. More red, but definitely not electric blue.

    Sometimes, if it is important for me to preserve the real shade of the sky, I climb into the editor and process the photo right on the spot, referring to live removals.

    It would be very easy to access the Camera app and use the slider to display the value to the one you want.

    Back to the yellow iPhone 14. Below are two highlights: the end result after processing top and them originals below.

    wp-image-1196659wp-image-1196662wp-image-1196660wp-image-1196661
    wp-image-1196652wp-image-1196646wp-image-1196649wp-image-1196653

    What I decided to add to the post is sustained plus or minus in the color scheme.

    In auto mode, the background, devices and covers all the time received particles. At the same time, unfortunately, tightening the white balance was not enough.

    For example, when the yellow color of the smartphone became payable in different photographs, for some reason the tree had a completely different shade. He preferred to align with masks, selective color correction, and so on.


    One static video, and we have 50 lying on white. Made for a webinar

    This problem gets even worse in video, when it comes to the hand in the frame, the background will be one color, and after another – Or when the dominant color in the frame changes, the reading algorithms instantly twist all other elements.

    The only way out is the light on the subject, but along with the color, the focus and exposure are also fixed, which immediately limits the flexibility of shooting.

    Why prevalence still not added? Over the past 10 years, Apple has automated color processing and differences in photos and videos too much. Half of all the work is now done not by iron cameras, but by the SmartHDR algorithm, which glues dozens of images into one frame.

    Sometimes, taking and reprogramming it to manually listen to the white balance is expensive and difficult.

    The problems don’t end there.

    Artificial sharpness annoys everyone already



    Above are examples of how, with artificial growth, algorithms turn into a drawing with paints.

    When I’m at homeC grade” (or telephoto) and the light is not strong enough, one of two things happens: instead of optically zooming, the iPhone decides to give scrap with the main camera or proximity module includes an aggressive noise suppression and work with Deep Fusion, thus you give a very blurry picture with thick white stripes in contrasting areas.


    iPhone with white stripes lifts up contrast

    I can’t accept optical 3rd zoom, I can’t random noise reduction, I can’t forbid the algorithm to harden the contrast.

    None of these barbaric methods can be used, no matter how “professional” smartphone is called.

    I don’t see the need to port the best quality radial illusion photo that spills out when the first frame is assembled. The text, for example, remains legible even without white halos, but otherwise it is terribly harmful from the point of view of aesthetics.

    Left photo before noise reduction, final direct shot

    See the rise above. Sometimes, after taking a picture, the iPhone is not going to process the photo in space-time, so when you open Galleryyou can see the grain in recent photographs.

    A matter of taste, of course, but I like this version of the frame much more than watercolor, which remains a second after processing by algorithms.

    It is clear that not every frame should remain like this. But that wouldn’t hurt. Sometimes a beautiful noise is better than a mess of feeling.

    Hope is near

    At the beginning of 2023, I saw an opinion on various channels and on the site that the iPhone 14 does not take good pictures. At least, than competitors younger than the same class.

    Many wrote that when shooting in 48 MP ProRAW mode on the iPhone 14 Pro, the extra contrast disappears. Text and other lines with a sharp transition of color begin to appear naturally, white color does not appear.

    Question from white balance more difficult.

    The iPad page for Final Cut Pro mentions the ability to “pro” use the tablet’s built-in camera.

    In addition to all the usual settings first Officially, Apple will be able to adjust the white balance.

    And although, judging by this preview, there will be no manual measurement, this is already a step forward. Apple in all 15 years of development of its cameras has never allowed to change the colors of the video balance before shooting.

    Now there is a lot of “marriage” from the iPhone

    Over the past three years, I have used three cameras of different calibers for work, one of which was an iPhone.

    Approximately half of all the photos in my articles are taken on it, and, in my opinion, the best of the mobile camera capabilities in the iPhone 13 Pro review a year later.

    That’s just behind the shots that arranged for me, are tens of thousands (!!!) photos that I would never be able to stretch to the quality that they wanted to be published here.


    How annoying are the neuronet white halos around the eyes

    A bad dancer is reluctant because it requires squeezing out the maximum of other types of skin, minor shadow correction, texture reduction, supersampling, adding appearance, and so on.

    But basic things like unpredictable white balance and unsolicited sharpness from AI add an incredible amount of work to the video and significantly spoil the result in the evening photo. We urgently need buttons to control them.

    Apple took small steps to prevent problems when it turned off oversharpening in 48MP ProRAW mode and made it possible to adjust the white balance when shooting video on the iPad through Final Cut Pro.

    A little bit of all this under the usual shooting on the iPhone, and the question is closed.

    We are waiting in the iPhone 15 Pro. Otherwise, it will be easier to take a good camera for plus or minus the same money and stay on the iPhone 13 Pro, which in all other respects is almost as good as the upcoming new items.

    To the point: Here is everything we expect from the iPhone 15. No fantasy

    📸 All photos in the article:






    Source: Iphones RU

    Previous articleElon Musk banned again from posting on social networks without legal approvalScience and technology09:37 | 16 May 2023
    Next articleIt’s not just everything: an expert from Russia says whether neural networks can predict the futurein Russia10:13 | 16 May 2023
    I am a professional journalist and content creator with extensive experience writing for news websites. I currently work as an author at Gadget Onus, where I specialize in covering hot news topics. My written pieces have been published on some of the biggest media outlets around the world, including The Guardian and BBC News.

    LEAVE A REPLY

    Please enter your comment!
    Please enter your name here