Adaptation of Cinematographic Techniques for Mobile and Vertical Video Formats

Abstract

The article describes how, in the context of the rapid growth of video consumption on mobile devices and the dominance of portrait screen orientation (9:16), classical cinematographic aspect ratios and techniques must be reconsidered and adapted. While traditional aspect ratios (4:3, 2.39:1) reflected historical technological and aesthetic constraints, the modern user experience dictates new rules of composition, blocking, editing, and technical implementation arising from the characteristics of touch controls and continuous scrolling. The study aims to identify and systematize the principles for applying classical cinematographic techniques in the mobile vertical format, to substantiate their effectiveness, and to develop a new grammar of visual storytelling for the narrow frame. The relevance of this work is driven by the unprecedented growth in mobile video traffic and changes in audience engagement models, which demand scholarly reflection on the aesthetic and technical transformations. Its novelty lies in the comprehensive integration of historical-technical analysis, ergonomic experiments, neurophysiological measurements, and content-analytic data to build a unified model of vertical video language. The methodological foundation combines comparative format analysis, ergonomic tests, a systematic review of recommendations on composition and camera movement, as well as empirical EEG-response studies and statistical analysis of audience behavior on TikTok and Reels. Results show that the central vertical axis of the frame serves as the path of least resistance for gaze and interaction; the rule of thirds transforms into vertical dynamics; and the Z-axis and dolly-in/out techniques become crucial for dramaturgy. Vertical split-screen significantly increases retention;gyrostabilized POV reduces fatigue; light gradients and pinpoint color accents guide attention from top to bottom; and binaural and tactile sound expand perception of the narrow frame. The study’s conclusions establish the theoretical and practical basis of a new cinematographic grammar for mobile vertical video: each modality—composition, camera movement, editing, lighting, color, sound, and interface effects—interacts to precisely direct attention within the 9:16 frame. This article will be particularly helpful to video production specialists, mobile app UX designers, and media aesthetics researchers.

Source type: Journals
Years of coverage from 2019
inLibrary
Google Scholar
HAC
doi
 

Downloads

Download data is not yet available.
To share
Herasymiuk Heorhii. (2025). Adaptation of Cinematographic Techniques for Mobile and Vertical Video Formats. The American Journal of Engineering and Technology, 7(8), 259–267. https://doi.org/10.37547/tajet/Volume07Issue08-21
Crossref
Сrossref
Scopus
Scopus

Abstract

The article describes how, in the context of the rapid growth of video consumption on mobile devices and the dominance of portrait screen orientation (9:16), classical cinematographic aspect ratios and techniques must be reconsidered and adapted. While traditional aspect ratios (4:3, 2.39:1) reflected historical technological and aesthetic constraints, the modern user experience dictates new rules of composition, blocking, editing, and technical implementation arising from the characteristics of touch controls and continuous scrolling. The study aims to identify and systematize the principles for applying classical cinematographic techniques in the mobile vertical format, to substantiate their effectiveness, and to develop a new grammar of visual storytelling for the narrow frame. The relevance of this work is driven by the unprecedented growth in mobile video traffic and changes in audience engagement models, which demand scholarly reflection on the aesthetic and technical transformations. Its novelty lies in the comprehensive integration of historical-technical analysis, ergonomic experiments, neurophysiological measurements, and content-analytic data to build a unified model of vertical video language. The methodological foundation combines comparative format analysis, ergonomic tests, a systematic review of recommendations on composition and camera movement, as well as empirical EEG-response studies and statistical analysis of audience behavior on TikTok and Reels. Results show that the central vertical axis of the frame serves as the path of least resistance for gaze and interaction; the rule of thirds transforms into vertical dynamics; and the Z-axis and dolly-in/out techniques become crucial for dramaturgy. Vertical split-screen significantly increases retention;gyrostabilized POV reduces fatigue; light gradients and pinpoint color accents guide attention from top to bottom; and binaural and tactile sound expand perception of the narrow frame. The study’s conclusions establish the theoretical and practical basis of a new cinematographic grammar for mobile vertical video: each modality—composition, camera movement, editing, lighting, color, sound, and interface effects—interacts to precisely direct attention within the 9:16 frame. This article will be particularly helpful to video production specialists, mobile app UX designers, and media aesthetics researchers.


background image

The American Journal of Engineering and Technology

259

https://www.theamericanjournals.com/index.php/tajet

TYPE

Original Research

PAGE NO.

259-267

DOI

10.37547/tajet/Volume07Issue08-21



OPEN ACCESS

SUBMITED

25 July 2025

ACCEPTED

07 August 2025

PUBLISHED

21 August 2025

VOLUME

Vol.07 Issue 08 2025

CITATION

Herasymiuk Heorhii. (2025). Adaptation of Cinematographic Techniques for
Mobile and Vertical Video Formats. The American Journal of Engineering and
Technology, 7(8), 259

267. https://doi.org/10.37547/tajet/Volume07Issue08-

21

COPYRIGHT

© 2025 Original content from this work may be used under the terms
of the creative commons attributes 4.0 License.

Adaptation of
Cinematographic
Techniques for Mobile and
Vertical Video Formats

Herasymiuk Heorhii

Videographer Miami, USA

Abstract:

The article describes how, in the context of

the rapid growth of video consumption on mobile
devices and the dominance of portrait screen
orientation (9:16), classical cinematographic aspect
ratios and techniques must be reconsidered and
adapted. While traditional aspect ratios (4:3, 2.39:1)
reflected historical technological and aesthetic
constraints, the modern user experience dictates new
rules of composition, blocking, editing, and technical
implementation arising from the characteristics of
touch controls and continuous scrolling. The study aims
to identify and systematize the principles for applying
classical cinematographic techniques in the mobile
vertical format, to substantiate their effectiveness, and
to develop a new grammar of visual storytelling for the
narrow frame. The relevance of this work is driven by
the unprecedented growth in mobile video traffic and
changes in audience engagement models, which
demand scholarly reflection on the aesthetic and
technical transformations. Its novelty lies in the
comprehensive integration of historical-technical
analysis, ergonomic experiments, neurophysiological
measurements, and content-analytic data to build a
unified model of vertical video language. The
methodological foundation combines comparative
format analysis, ergonomic tests, a systematic review of
recommendations on composition and camera
movement, as well as empirical EEG-response studies
and statistical analysis of audience behavior on TikTok
and Reels. Results show that the central vertical axis of
the frame serves as the path of least resistance for gaze
and interaction; the rule of thirds transforms into
vertical dynamics; and the Z-axis and dolly-in/out
techniques become crucial for dramaturgy. Vertical
split-screen significantly increases retention;


background image

The American Journal of Engineering and Technology

260

https://www.theamericanjournals.com/index.php/tajet

The American Journal of Engineering and Technology

gyrostabilized POV reduces fatigue; light gradients and
pinpoint color accents guide attention from top to
bottom; and binaural and tactile sound expand

perception of the narrow frame. The study’s conclusions

establish the theoretical and practical basis of a new
cinematographic grammar for mobile vertical video:
each modality

composition, camera movement,

editing, lighting, color, sound, and interface effects

interacts to precisely direct attention within the 9:16
frame. This article will be particularly helpful to video
production specialists, mobile app UX designers, and
media aesthetics researchers.

Keywords:

adaptation; cinematographic techniques;

vertical format; mobile video; composition; editing;
stabilization; lighting; color; sound

Introduction

Over the past five years, the mobile screen has ceased
to be secondary and has become the primary media
carrier. According to Vidico, 75% of all video views now
occur on smartphones, which users hold vertically
almost all the time, rather than rotating them for a
horizontal frame [1]. The shift is also visible in everyday
media diets: the British TouchPoints-2025 survey
recorded that adults spend 3h 21min per day on their
phones

for the first time exceeding time spent

watching linear television (3 h 16 min)

and do so

evenly throughout the day, whereas TV viewing still
peaks in the evening [2].

Such a redistribution of attention is supported by
network infrastructure: Ericsson estimates that by the
end of 2024, video applications will have generated 111
exabytes of the 147 exabytes of monthly smartphone
traffic, i.e., over three-quarters of the total data volume,
with a six-year growth exceeding sevenfold [3]. In
practical terms, this means that the energy consumption
of mobile video streams alone over one month is

comparable to London’s five

-month needs, underlining

not only the cultural but also the infrastructural
significance of this format.

Against this backdrop, the evolution of cinematographic
aspect ratios takes on a new dimension. Early film stock
and the 4:3 television standard

formally codified as the

Academy ratio of 1.37:1 in 1932

provided visual

continuity with the theatrical projection window and the
technological limits of the first sound track [4]. In the
1950s, in response to the challenge posed by television,
studios transitioned to anamorphic CinemaScope. They

established the 2.39:1 wide frame, which remains one of
the standard screen ratios in digital cinema distribution.

However, the logic of the wider, the more cinematic no
longer applies in an environment dominated by portrait
device orientation. Snapchat was the first to
demonstrate this: as early as the mid-2010s, the
company reported that vertical advertisements were
watched to completion more often than horizontal
ones, provided users were not required to rotate their
screens [5].

With the advent of TikTok and short Reels, the vertical
frame ceased to be merely a cropped version of a film.
It became a new foundational 9:16 stock, dictating its
own rules of composition, blocking, and pacing, and
demanding scholarly reflection on the aesthetic and
technological changes at the intersection of mobile UX
and classical cinematography.

Materials and Methodology

This study is based on the analysis of 17 key sources,
including academic articles, industry reports, empirical
experiments, and user cases. The theoretical foundation
comprises

studies

of

consumer

habits

and

infrastructural trends. The Vidico report [1] and the
TouchPoints-2025 survey [2] revealed the dominance of
vertical viewing and the even distribution of mobile
consumption throughout the day. The DataReportal
report [3] confirmed the rapid growth of video-app
traffic. The historical-technical background of pre-
cinematographic aspect ratios was provided by SMPTE
standards [4], and practical conclusions on the
effectiveness of the vertical format were drawn from the
Snapchat case [5]. Compositional and ergonomic
aspects were drawn from classic methodologies by Perry
and Hourcade [6] and modern vertical framing
recommendations by Parisi [7], supplemented with
theoretical insights on split-screen engagement metrics
by Clayton [8] and Spyrou [9].

Methodologically, the research combines several
approaches. A comparative format analysis juxtaposed
classical cinematic aspect ratios (4:3, 2.39:1) with the
vertical 9:16, considering historical constraints and
current

user

preferences

[4,

5].

Ergonomic

experiments

conducted following Perry and Hourcade

[6] and Parisi [7]

demonstrated the path of least

resistance along the vertical screen axis and the
transformation of the rule of thirds in the vertical
format.

A

systematic

review

of

theoretical


background image

The American Journal of Engineering and Technology

261

https://www.theamericanjournals.com/index.php/tajet

The American Journal of Engineering and Technology

recommendations encompassed studies on the
influence of shooting angle on perceived trust
(Baranowski & Hecht [10]), the effects of various
camera-movement techniques (Yilmaz et al. [11]), and
the algorithmic aspects of video stabilization (Luchetti et
al. [13]).

Content and statistical analyses included: investigation

of duet and reaction videos’ impact on audience

retention in TikTok/Reels (Spyrou [9]); analysis of the
pocket-gimbal market

sales forecasts and CAGR

through 2033 (Archive Market Research [12]);
neurophysiological EEG measurements of principal
components in experiments by Huttunen [14] and
Mattavelli et al. [15] to assess emotional and cognitive
responses to lighting schemes; and quantitative analysis
of color accents in mobile feeds

click-through rates for

warm palettes (Sokolik et al. [16]) and the U-shaped
relationship between saturation and conversion (Wen et
al. [17]). The integration of these methods provided a
comprehensive picture of how the vertical format
reconfigures composition, camera movement, and
technical solutions in the mobile context.

Results and Discussion

Reorienting the frame from landscape to portrait has led
to a reassembly of the classical compositional grid.
Ergonomic studies of touch input show that, when
holding the device with one hand, users tap elements
located in the central band of the screen more quickly
and accurately. In contrast, objects at the top and
bottom edges require greater joint movement

this

reinforces the value of the longitudinal axis as a path of
least resistance for gaze and thumb interaction [6].

At the same time, the rule of thirds does not disappear;
instead, it transforms: instead of the familiar
displacement of the protagonist to the left or right
intersection point, attention is distributed along the
vertical axis. Creators of vertical videos recommend
placing the dominant object within the middle third of
the width, yet smoothly sliding it between the upper and

lower horizontal dividers to engage the eye’s natural

scroll movement and preserve interface symmetry [7].

The limited horizontal span necessitates a reevaluation
of spatial relationships within the frame: horizontal
mise-en-scène gives way to depth staging. The vertical
format more easily accommodates a full-length human
figure, enabling storytelling through movement toward
or away from the camera rather than lateral tracking

shots. Theorists of vertical cinema note that it is the Z-
axis, not the X-axis, that becomes the principal carrier of
dramaturgy: the foreground jumps interactively toward
the viewer, the midground sustains the action, and the
background provides context

forming a three-layer

column instead of a panoramic horizon [8].

The lack of width is compensated for by multi-level
editing. The most popular technique is the vertical split-

screen. TikTok’s and Reels’ algorithms rank duets and

reaction videos higher because the simultaneous
stimulation of two screen regions increases average

watch time by ≈ approximately 40% and boosts

information retention by up to 60%, thanks to
concurrent visual and auditory channels [9]. However,
cognitive experiments also record increased working-
memory load. With an excess of parallel tracks, users
reproduce the core message less accurately, demanding
a strict hierarchy of layers and a contrasting rhythm of
shot changes.

Thus, the central axis establishes a fulcrum for rapid
perception, depth becomes the primary tool for plot
construction, and the multi-level frame expands

perceived space without widening the screen’s

boundaries. These three principles operate in concert
and form the basis of a cinematographic grammar for
the vertical 9:16 format.

When the vertical format fixes the gaze on a narrow axis,
camera movement assumes the function of expanding
space and redistributing semantic emphasis. Vertical
pans and classic tilt-up/tilt-down movements prove
especially effective, as they scan the frame along the

same vector in which the viewer’s finger scrolls the feed.

Here, filmic language merges with gesture navigation: a

camera rising from the floor to the subject’s face shifts
attention from the interface’s comments zone to the

protagonist. At the same time, the reverse movement
prepares the viewer for the next swipe. Simultaneously,
the vertical angle functions as a semantic filter: an
experiment [10] demonstrated that even a 30° tilt of the
camera from above or below alters trust ratings for the
character; audiences ascribed maximum reliability to
eye-level shots, whereas a high angle statistically
reduced trust scores by 0.8 points on a nine-point scale,
as shown in Fig. 1. Thus, tilt-up

as a marker of status

elevation

and tilt-down

as a distancing effect

become formal antonyms within the same portrait
geometry.


background image

The American Journal of Engineering and Technology

262

https://www.theamericanjournals.com/index.php/tajet

The American Journal of Engineering and Technology

Fig. 1. Ratings for trust by camera angle and sex [10]

Depth movement along the Z-axis, which in horizontal
cinema was supported by complex multi-plane mise-en-
scène, is simplified in the vertical format to dolly-in and
dolly-out. When shooting on a smartphone, these
techniques visually draw the viewer closer to or farther
from the event, allowing one to experience approach
and withdrawal without lateral displacement. A
neurocinematographic study with 44 respondents
compared dolly, steadicam, handheld camera, and static
shots. Smooth forward movement yielded a median
immersion index higher than static shots, although
emotional valence remained unchanged [11]. For the
mobile viewer, this means that dolly-in can serve as a
universal emotional accent. By minimizing editing,

creators achieve significant engagement gains through
pure kinetics, along the same vertical tunnel that houses
the protagonist.

The third component of vertical cinematographic
grammar is gyrostabilized POV. The small amplitude of
micro-vibrations, barely perceptible on a cinema screen,
becomes visual noise on a six-inch display, increasing the
risk of acute fatigue and motion sickness. Consequently,
demand for pocket three-axis stabilizers is growing
faster than the mobile-video market itself; an industry
report estimates global sales of such devices at USD 1.5
billion in 2025, with a projected CAGR of 15% through
2033, as illustrated in Fig. 2 [12].


background image

The American Journal of Engineering and Technology

263

https://www.theamericanjournals.com/index.php/tajet

The American Journal of Engineering and Technology

Fig. 2. Smartphone Gimbal Stabilizers Market Size [12]

The stabilization effect is measured not only
commercially, but also in other ways. In a sample
subjected to the standard Simulator Sickness
Questionnaire, the subjective severity of symptoms
after viewing the same clips, but with offline trajectory
stabilization, was statistically lower than the baseline
level [13]. Thus, gyrostabilized POV simultaneously
enhances physiological comfort and preserves presence
through slight correction of hand-held camera motion,
making it the primary technical solution for mobile
narrative.

Together, vertical pans, depth-axis dolly passes, and
stabilized POV create a virtual volume in the narrow
frame: the first establishes a hierarchy along the Y-axis,
the second adds emotional rhythm along the Z-axis, and
the third holds the viewer within this structure without
kinetic noise. In subsequent sections, it will be shown
how light, color, and sound amplify this kinetic matrix,
turning the 9:16 constraint into a tool for precise
direction of attention.

In the vertical frame, light becomes not only an exposure
tool but also a compositional one: a gradient along the

Y-axis replaces missing width. A classic Rembrandt-style
technique

placing the key light high above the lens and

slightly to one side, with a fill light below

creates a

sliding vertical transition from a bright forehead to a
darker torso. Such a gradient structures the scene much
like the tilt-up discussed previously: the eye first reads
the upper headline details, then naturally descends to
the action while the background remains in mid-tone
and does not compete with the app interface. Portrait
photography has long demonstrated that an
asymmetrical diagonal of light in the upper quarter of
the frame renders the face both three-dimensional and
clickable, and requires only one lamp and reflector,
critical for on-the-move mobile shooting.

Neurophysiological data reinforce this technological
logic. In a pilot EEG experiment where faces were lit
from above, below, and in silhouette, underlighting that
slipped beneath the lower-panel icons

and complete

eye-region darkness

elicited a more negative early

posterior negativity than the classic 45° top key,
interpreted as a subconscious vigilance signal [14], as
illustrated in Fig. 3.


background image

The American Journal of Engineering and Technology

264

https://www.theamericanjournals.com/index.php/tajet

The American Journal of Engineering and Technology

Fig. 3. The mean signal amplitudes elicited by underlight and 45-degree light [14]

The graph shows a noticeable difference in the mean
negativity within both the 150

300 and 200

600 ms

after-stimulus intervals, although the latter did not
reach statistical significance. This suggests that lighting
that originates from below the face, distorting the facial
features, elicits a stronger early emotional response
than lighting that reveals the entire face.

In other words, an error in vertical-contrast distribution

is instantaneously reflected in the viewer’s emotional

baseline before the narrative even begins. A
psychophysical study [15] confirmed the practical
finding: with a central gaze model, trust indices
statistically decreased under a split-light condition (a
harsh horizontal shadow), whereas a moderate top-
down Rembrandt gradient increased perceptions of
trustworthiness and attractiveness, especially in
younger faces.

The vertical gradient also applies to color. In an endless
scroll, attention is captured by small spectral bursts

localized but not aggressive accents. A field sample of
1.5 million impressions showed that a warm red palette
in a narrow vertical box generated clicks significantly
more often than a cool blue one, confirming that a warm
hue shouts from a monotone feed [16]. A more recent
machine analysis of 2,984 short videos revealed a U-

shaped relationship: moderate or very high saturation
increased subsequent purchases, whereas mid-range
values were lost against neighboring clips [17]. Thus, for
the mobile-vertical format, the point-color rule applies:
a bright accent in one third of the screen amplifies the
dopaminergic hook in the first milliseconds, but
excessive colorfulness disintegrates into noise and
reduces retention.

Altogether, a top-to-bottom light gradient, calibrated

facial contrast, and measured color guide the viewer’s

gaze along the same path as the camera movement
described earlier: from above down to the focal point
and then toward the interface zone. This transforms the
9:16 limitation into a multi-layered channel of directed
sensory information, where lighting not only paints the

subject but also synchronizes the viewer’s physiological

rhythm with the edit pace of the clip.

Sound in vertical video functions as an invisible
extension of the frame, filling what is bounded by the
narrow screen. Spatial, binaural, recording creates the
illusion of volume: the viewer perceives footsteps
behind or a voice to the side, even though only a portrait
plane is visible. Through such acoustic completion, the
director is freed from constantly changing camera
angles, and the audience receives a convincing three-


background image

The American Journal of Engineering and Technology

265

https://www.theamericanjournals.com/index.php/tajet

The American Journal of Engineering and Technology

dimensional impression even with a static camera. The
smooth movement of the sound source around the head
directs the movement of attention, preparing the viewer
for the appearance of an object in the center of the
frame.

The vertical experience is amplified not only auditorily
but also somatically. Low frequencies, transmitted

through the speaker or via vibration of the phone’s

div, create a tactile background that underscores the
scale or emotional weight of a scene. A brief bass
impulse reinforces the close-up of the protagonist, while
prolonged sub-bass masks location changes by
smoothing abrupt edits. The user perceives a light
pulsing pressure in the palm, transforming the spectacle
into a kinesthetic event.

Editing in the mobile context relies on the speed of
perception. Retention is aided by a sharp hook: the
opening moments present the primary visual stimulus,
followed by a rapid sequence of short shots. The rhythm
is calibrated so that each shot delivers a fresh piece of
information but leaves no room for the viewer to skip to
the following clip. Whereas traditional editing revolved
around the complete thought of a scene, here it is
oriented toward micro-transitions that integrate into
the endless feed and synchronize with the scroll gesture.

Text elements adhere to the same principle of time
economy. Titles do not appear as separate cards but as
animated typography embedded in the trajectory of the
camera or finger movement. Letters grow from the
horizon line, assemble, and disappear at the moment
the visual emphasis shifts. Such living text does not
hijack attention for long nor obscure the face, yet it
registers in memory as part of the unified motion.

An additional layer of engagement is created by
augmented reality. Filters that respond to facial
expressions or gestures connect the narrative to the

viewer: a mask that mirrors one’s expression or an

interactive object caught by a swipe transform passive
viewing into a personal event. A similar effect is
achieved by the comic-panel composition, where the
main vertical frame remains fixed while auxiliary layers
scroll independently, revealing alternative angles or plot
details.

Together, spatial sound, tactile bass, accelerated
editing, kinetic typography, and the AR layer form a
synesthetic system in which each channel supports the

screen’s confined geometry. The light gradients

described earlier map the visual route from top to
bottom; audio markers and vibration anchor the stages
of this route; visual effects hold attention on key points.
Thus, the vertical frame becomes not a limitation but a
platform where all modalities work in concert to direct
the viewer to the center of the story.

Conclusion

The conducted study demonstrates that the shift to the
mobile vertical 9:16 video format has triggered a
comprehensive set of aesthetic and technical
transformations of classical cinematographic language.
Within the narrow frame, the central longitudinal axis
serves as the path of least resistance for gaze and
interface interaction, defining a new compositional
paradigm. Instead of the habitual horizontal distribution
of emphasis, we observe a vertical zone of interest in
which positioning the dominant object within the
middle third of the width and dynamically sliding it along

the frame’s height harmonize with the user’s natural

scroll, facilitating perception and enhancing attention
retention.

The depth Z-axis becomes the primary dramaturgical
tool: dolly-in and dolly-out replace complex multi-plane
compositions, creating the effect of pushing the object
toward the viewer or detaching it from the background.
Simultaneously, multi-level editing

such as vertical

split-screen

—expands the narrow frame’s perceived

space without losing image coherence, stimulating
visual and auditory perception through parallel
information streams.

Technical solutions, such as gyrostabilized POV, ensure
the necessary smoothness of movement despite the
limitations of handheld smartphone shooting, thereby
minimizing visual noise and reducing the risk of fatigue.
The combination of vertical pans, axial movements, and
stabilization forms a virtual volume within the portrait
frame, in which lighting, color, and sound perform not
only utilitarian but also compositional functions.
Directed light gradients structure perception from top to
bottom; point-color accents shout from the endless
feed; and binaural and tactile components complement
the visual track, creating a synesthetic system of
interaction.

Thus, the adaptation of classical cinematographic
techniques to the 9:16 format goes beyond a mere
rotation of the frame and constructs a new grammar of
mobile narrative, in which each modality

from


background image

The American Journal of Engineering and Technology

266

https://www.theamericanjournals.com/index.php/tajet

The American Journal of Engineering and Technology

composition and editing to lighting, color, sound, and
interface effects

acts in concert, turning the vertical

screen constraint into a tool for precise attention
direction. These conclusions provide the theoretical and
practical foundation for the further development of
vertical film language and the optimization of user
experience in mobile video applications.

References

1.

“20 Amazing Video Consumption Trends To Watch
Out

For,”

Vidico,

Aug.

29,

2024.

https://vidico.com/news/video-consumption-
trends/ (accessed May 26, 2025).

2.

M. Sweney, “Adults in Great Britain now spending
more time on mobiles than watching TV,” The

Guardian,

2025.

https://www.theguardian.com/media/2025/jun/25
/adults-great-britain-time-mobiles-watching-tv-
screen-ipa-survey (accessed May 27, 2025).

3.

“Digital 2025: Mobile Data Consumption Trends,”

DataReportal,

Feb.

05,

2025.

https://datareportal.com/reports/digital-2025-sub-
section-mobile-data-consumption (accessed May
28, 2025).

4.

“SMPTE

ST

430

-

16:2017,”

SMPTE,

2017.

https://pub.smpte.org/doc/st430-16/20170714-
pub/st0430-16-2017.pdf (accessed May 29, 2025).

5.

S. Fatah, “Snapchat launches new program,
highlights impact of vertical video,” Media in

Canada,

2016.

https://mediaincanada.com/2016/06/14/snapchat-
launches-new-program-highlights-impact-of-
vertical-video/ (accessed May 30, 2025).

6.

K. B. Perry and J. P. Hourcade, “Evaluating one

-

handed thumb tapping on mobile touchscreen

devices,” Graphics Interface, pp. 57–

64, May 2008,

Accessed: Jun. 26, 2025. [Online]. Available:
https://www.researchgate.net/publication/221474
717_Evaluating_one_handed_thumb_tapping_on_
mobile_touchscreen_devices

7.

D. Parisi, “Vertical Video for Social Media: Best
Practices for Reels in 2025,” LocalEyes, Apr. 08,

2025.

https://localeyesit.com/vertical-video-for-

social-media/ (accessed Jun. 01, 2025).

8.

R. Clayton, “This is a repository copy of Filmmaking
Theory for Vertical Video Production,” University of

Leeds, 2019. Accessed: Jun. 01, 2025. [Online].

Available:
https://eprints.whiterose.ac.uk/id/eprint/180677/1
/EuroMedia2019_52556.pdf

9.

Stéphanos Spyrou, “Discover how split

-screen

videos on TikTok are doubling engagement rates,”

LinkedIn,

Dec.

03,

2024.

https://www.linkedin.com/pulse/split-screen-
phenomenon-double-content-distraction-
st%C3%A9phanos-spyrou-bbg3e (accessed Jun. 02,
2025).

10.

A. M. Baranowski and H. Hecht, “Effect of Camera
Angle on Perception of Trust and Attractiveness,”

Empirical Studies of the Arts, vol. 36, no. 1, pp. 90

100,

May

2017.

https://experimental.psychologie.uni-
mainz.de/files/2024/06/2018BaranowskiHecht_Ca
mera_Angle.pdf

(accessed Jun. 02, 2025).

11.

M. B. Yilmaz, E. Lotman, A. Karjus, and P. Tikka, “An

embodiment of the cinematographer: emotional
and perceptual responses to different camera

movement techniques,” ProQuest, vol. 17, Jul. 2023,

doi: https://doi.org/10.3389/fnins.2023.1160843.

12.

“Smartphone Gimbal Stabilizers Market Size,”

Archive

market

research,

2024.

https://www.archivemarketresearch.com/reports/
smartphone-gimbal-stabilizer-260840#summary
(accessed Jun. 05, 2025).

13.

A. Luchetti, M. Zanetti, D. Kalkofen, and M. D. Cecco,

“Stabilization of spherical videos based on feature
uncertainty,” The Visual Computer, vol. 39, no. 9,

pp.

4103

4116,

Jul.

2022,

doi:

https://doi.org/10.1007/s00371-022-02578-z.

14.

S. Huttunen, “Faces in shadows: silhouette light,

underlight and toplight elicit increased early

posterior negativity,” Frontiers in Neuroscience, vol.

19,

Mar.

2025,

doi:

https://doi.org/10.3389/fnins.2025.1553977.

15.

G. Mattavelli, I. Gorrino, E. Cesana, J. D. Angelis, and

P. Ricciardelli, “Illumination and gaze effects on face

evaluation: The Bi-

AGI database,” Frontiers in

Psychology,

vol.

13,

Oct.

2022,

doi:

https://doi.org/10.3389/fpsyg.2022.948142.

16.

K. Sokolik, R. G. Magee, and J. D. Ivory, “Red

-Hot and

Ice-

Cold Web Ads: The Influence of Web Ads’ Warm

and Cool Colors on Click-

Through Rates,” Journal of

Interactive Advertising, vol. 14, no. 1, pp. 31

37, Jan.


background image

The American Journal of Engineering and Technology

267

https://www.theamericanjournals.com/index.php/tajet

The American Journal of Engineering and Technology

2014,

doi:

https://doi.org/10.1080/15252019.2014.907757

.

17.

X. Wen, Y. Niu, and C. Liao, “U

-shaped Effect of Color

in Short-form Video: Combination of Text, Audio,

Image ML and Generative AI,” Journal of Business

Research, vol. 192, pp. 115278

115278, Mar. 2025,

doi:
https://doi.org/10.1016/j.jbusres.2025.115278.

References

“20 Amazing Video Consumption Trends To Watch Out For,” Vidico, Aug. 29, 2024. https://vidico.com/news/video-consumption-trends/ (accessed May 26, 2025).

M. Sweney, “Adults in Great Britain now spending more time on mobiles than watching TV,” The Guardian, 2025. https://www.theguardian.com/media/2025/jun/25/adults-great-britain-time-mobiles-watching-tv-screen-ipa-survey (accessed May 27, 2025).

“Digital 2025: Mobile Data Consumption Trends,” DataReportal, Feb. 05, 2025. https://datareportal.com/reports/digital-2025-sub-section-mobile-data-consumption (accessed May 28, 2025).

“SMPTE ST 430-16:2017,” SMPTE, 2017. https://pub.smpte.org/doc/st430-16/20170714-pub/st0430-16-2017.pdf (accessed May 29, 2025).

S. Fatah, “Snapchat launches new program, highlights impact of vertical video,” Media in Canada, 2016. https://mediaincanada.com/2016/06/14/snapchat-launches-new-program-highlights-impact-of-vertical-video/ (accessed May 30, 2025).

K. B. Perry and J. P. Hourcade, “Evaluating one-handed thumb tapping on mobile touchscreen devices,” Graphics Interface, pp. 57–64, May 2008, Accessed: Jun. 26, 2025. [Online]. Available: https://www.researchgate.net/publication/221474717_Evaluating_one_handed_thumb_tapping_on_mobile_touchscreen_devices

D. Parisi, “Vertical Video for Social Media: Best Practices for Reels in 2025,” LocalEyes, Apr. 08, 2025. https://localeyesit.com/vertical-video-for-social-media/ (accessed Jun. 01, 2025).

R. Clayton, “This is a repository copy of Filmmaking Theory for Vertical Video Production,” University of Leeds, 2019. Accessed: Jun. 01, 2025. [Online]. Available: https://eprints.whiterose.ac.uk/id/eprint/180677/1/EuroMedia2019_52556.pdf

Stéphanos Spyrou, “Discover how split-screen videos on TikTok are doubling engagement rates,” LinkedIn, Dec. 03, 2024. https://www.linkedin.com/pulse/split-screen-phenomenon-double-content-distraction-st%C3%A9phanos-spyrou-bbg3e (accessed Jun. 02, 2025).

A. M. Baranowski and H. Hecht, “Effect of Camera Angle on Perception of Trust and Attractiveness,” Empirical Studies of the Arts, vol. 36, no. 1, pp. 90–100, May 2017. https://experimental.psychologie.uni-mainz.de/files/2024/06/2018BaranowskiHecht_Camera_Angle.pdf (accessed Jun. 02, 2025).

M. B. Yilmaz, E. Lotman, A. Karjus, and P. Tikka, “An embodiment of the cinematographer: emotional and perceptual responses to different camera movement techniques,” ProQuest, vol. 17, Jul. 2023, doi: https://doi.org/10.3389/fnins.2023.1160843.

“Smartphone Gimbal Stabilizers Market Size,” Archive market research, 2024. https://www.archivemarketresearch.com/reports/smartphone-gimbal-stabilizer-260840#summary (accessed Jun. 05, 2025).

A. Luchetti, M. Zanetti, D. Kalkofen, and M. D. Cecco, “Stabilization of spherical videos based on feature uncertainty,” The Visual Computer, vol. 39, no. 9, pp. 4103–4116, Jul. 2022, doi: https://doi.org/10.1007/s00371-022-02578-z.

S. Huttunen, “Faces in shadows: silhouette light, underlight and toplight elicit increased early posterior negativity,” Frontiers in Neuroscience, vol. 19, Mar. 2025, doi: https://doi.org/10.3389/fnins.2025.1553977.

G. Mattavelli, I. Gorrino, E. Cesana, J. D. Angelis, and P. Ricciardelli, “Illumination and gaze effects on face evaluation: The Bi-AGI database,” Frontiers in Psychology, vol. 13, Oct. 2022, doi: https://doi.org/10.3389/fpsyg.2022.948142.

K. Sokolik, R. G. Magee, and J. D. Ivory, “Red-Hot and Ice-Cold Web Ads: The Influence of Web Ads’ Warm and Cool Colors on Click-Through Rates,” Journal of Interactive Advertising, vol. 14, no. 1, pp. 31–37, Jan. 2014, doi: https://doi.org/10.1080/15252019.2014.907757.

X. Wen, Y. Niu, and C. Liao, “U-shaped Effect of Color in Short-form Video: Combination of Text, Audio, Image ML and Generative AI,” Journal of Business Research, vol. 192, pp. 115278–115278, Mar. 2025, doi: https://doi.org/10.1016/j.jbusres.2025.115278.