Compare commits
183 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
db7700a6b9 | ||
|
|
222234f978 | ||
|
|
3672474ff5 | ||
|
|
5ff0fc3fad | ||
|
|
554ca4cc03 | ||
|
|
51febdfcc0 | ||
|
|
0588f47837 | ||
|
|
8b036af47f | ||
|
|
c66d4c2568 | ||
|
|
59dbb8985a | ||
|
|
5e57e57ad2 | ||
|
|
739acb0dd8 | ||
|
|
e9e535044e | ||
|
|
bc62801949 | ||
|
|
e2b6a4bf7c | ||
|
|
4716c4d11c | ||
|
|
871c22d8e8 | ||
|
|
e44fc93b4e | ||
|
|
71943dbabc | ||
|
|
0a026afba4 | ||
|
|
a4b3db3eba | ||
|
|
06978e6862 | ||
|
|
457b599188 | ||
|
|
125aeca340 | ||
|
|
1caf4beb72 | ||
|
|
49c1164b8b | ||
|
|
0f635b8a86 | ||
|
|
4b9f9a0364 | ||
|
|
2f3658de5b | ||
|
|
ec4af18e7a | ||
|
|
58b01f2be7 | ||
|
|
1d4507782b | ||
|
|
8c7eee580d | ||
|
|
303fd4bc80 | ||
|
|
5febb96916 | ||
|
|
b16e76370b | ||
|
|
feb5441251 | ||
|
|
ea182d4ddb | ||
|
|
f853cf0f85 | ||
|
|
b492be227a | ||
|
|
7fe5b66c0c | ||
|
|
07cc0cd95e | ||
|
|
de2d7c0593 | ||
|
|
826677cb03 | ||
|
|
95aeacf694 | ||
|
|
a3bb16e850 | ||
|
|
0ed85fce4a | ||
|
|
1a0a5f4482 | ||
|
|
06f6322d32 | ||
|
|
0cbcf1a702 | ||
|
|
f94310fdb7 | ||
|
|
06b523f3e8 | ||
|
|
efb4fbfc95 | ||
|
|
2abda01fe6 | ||
|
|
f007ada29f | ||
|
|
44916bf062 | ||
|
|
89129ae5c4 | ||
|
|
696b2b56d3 | ||
|
|
b9185b5b07 | ||
|
|
de0f4a57c1 | ||
|
|
207472283b | ||
|
|
22f4b00e76 | ||
|
|
95d858b2c6 | ||
| 3f0efab49b | |||
| 5c47f193d4 | |||
| 324084c845 | |||
|
|
79f088a86a | ||
|
|
9e37ff18c4 | ||
|
|
3647b25b4c | ||
|
|
772c1d8f90 | ||
|
|
ad58ba5ce6 | ||
|
|
24d0700db2 | ||
|
|
3463c1e371 | ||
|
|
5ca7d6d12c | ||
|
|
77dfb4b1d3 | ||
| 1df2e74566 | |||
| 8f9f77e891 | |||
| 6a03d4d6e2 | |||
| a263c735aa | |||
| 5e0fdd6476 | |||
| 1cfa51f2ca | |||
| a3dc894ba7 | |||
| c0305ec0a7 | |||
| 00ca7b92c1 | |||
| c4b3dd94f9 | |||
| 5307eda92e | |||
| 1620638110 | |||
| 467977c81c | |||
| fee7940660 | |||
| c03d4389ae | |||
| abdf13142e | |||
| 5c96439fa8 | |||
| 45a1c5aa4e | |||
| 42f9486c64 | |||
| 1eecd5a4f8 | |||
| c83f17dd44 | |||
| a01e8bfca5 | |||
| cf49ff06d1 | |||
| e3115e7557 | |||
| ce2f3993e1 | |||
| ee31634b0b | |||
| 30d22892f8 | |||
|
|
260c605201 | ||
| 747ff41ad3 | |||
| fe1ed57758 | |||
| a082058ce2 | |||
|
|
9739f9aee4 | ||
|
|
5bb7dcc795 | ||
| ec3fb25c7b | |||
| 9fee9d1ae4 | |||
|
|
e1cff6c8db | ||
|
|
93cc8a23c9 | ||
|
|
bcb4e4d3d6 | ||
|
|
dba494b4e6 | ||
|
|
ca57e981a6 | ||
|
|
ff93875a07 | ||
|
|
bc3b593362 | ||
|
|
f920dec475 | ||
|
|
84b1135861 | ||
|
|
aaa6b2cabc | ||
|
|
aea8c7e9ea | ||
|
|
b9aed1f0b6 | ||
|
|
12689fe61d | ||
|
|
a280248f6a | ||
|
|
6b99091a55 | ||
|
|
8ed6809ad0 | ||
|
|
882d021bb6 | ||
|
|
131cca2c53 | ||
|
|
a03449a32b | ||
|
|
7d7e43c6f0 | ||
|
|
1c9f67e47a | ||
|
|
eaee3b34da | ||
|
|
5e017a8373 | ||
| 123d8659e1 | |||
| 82b257d809 | |||
|
|
8ec7f9c2d1 | ||
|
|
7cc6efb9f1 | ||
|
|
87ccb7e8a6 | ||
|
|
3765f25fd8 | ||
|
|
74dfbe30d7 | ||
|
|
84baeb2d87 | ||
|
|
9df5973676 | ||
|
|
b492ebdab9 | ||
|
|
1ae52399b9 | ||
| 3008d66dfe | |||
| 9f22c70e89 | |||
| a5d568ba34 | |||
| 73c957c9bb | |||
| 7c899e32bb | |||
| e325aaa529 | |||
| fdc8f8f602 | |||
| 7fc025821b | |||
|
|
322321b1ed | ||
|
|
a46a2b421e | ||
|
|
0cda6390cd | ||
|
|
c963c2c675 | ||
|
|
37786f56b5 | ||
|
|
24e85d7005 | ||
|
|
9b57007d5e | ||
|
|
59503b566c | ||
|
|
b390eaf2a0 | ||
|
|
dcfc0f734b | ||
|
|
ab0df0fed3 | ||
|
|
8ad50fd881 | ||
|
|
5ed97e0b9b | ||
|
|
7d85b45d68 | ||
|
|
de56e85d2e | ||
| d84797e6a5 | |||
|
|
e3964b0002 | ||
|
|
54a3df7f7c | ||
|
|
f381fad31f | ||
|
|
e6734cb4ef | ||
|
|
dcd79b74fd | ||
|
|
4d7f728f25 | ||
|
|
ad7562f387 | ||
| e7426398ee | |||
| 20ae4e763f | |||
| 6fdcd20176 | |||
| ab864e472b | |||
|
|
45ba438c3d | ||
| 84d7a9b14b | |||
|
|
494ae89034 | ||
|
|
0a8d3a8f85 |
10
.gitignore
vendored
Normal file
10
.gitignore
vendored
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
__pycache__
|
||||||
|
junk/
|
||||||
|
.vscode
|
||||||
|
.ipynb_checkpoints/
|
||||||
|
ansible/inventory/hawaii.yml
|
||||||
|
ansible/inventory/peppermint.yml
|
||||||
|
ffx_test_report.log
|
||||||
|
bin/conversiontest.py
|
||||||
|
*.egg-info/
|
||||||
|
|
||||||
595
LICENSE.md
Normal file
595
LICENSE.md
Normal file
@@ -0,0 +1,595 @@
|
|||||||
|
GNU General Public License
|
||||||
|
==========================
|
||||||
|
|
||||||
|
_Version 3, 29 June 2007_
|
||||||
|
_Copyright © 2007 Free Software Foundation, Inc. <<http://fsf.org/>>_
|
||||||
|
|
||||||
|
Everyone is permitted to copy and distribute verbatim copies of this license
|
||||||
|
document, but changing it is not allowed.
|
||||||
|
|
||||||
|
## Preamble
|
||||||
|
|
||||||
|
The GNU General Public License is a free, copyleft license for software and other
|
||||||
|
kinds of works.
|
||||||
|
|
||||||
|
The licenses for most software and other practical works are designed to take away
|
||||||
|
your freedom to share and change the works. By contrast, the GNU General Public
|
||||||
|
License is intended to guarantee your freedom to share and change all versions of a
|
||||||
|
program--to make sure it remains free software for all its users. We, the Free
|
||||||
|
Software Foundation, use the GNU General Public License for most of our software; it
|
||||||
|
applies also to any other work released this way by its authors. You can apply it to
|
||||||
|
your programs, too.
|
||||||
|
|
||||||
|
When we speak of free software, we are referring to freedom, not price. Our General
|
||||||
|
Public Licenses are designed to make sure that you have the freedom to distribute
|
||||||
|
copies of free software (and charge for them if you wish), that you receive source
|
||||||
|
code or can get it if you want it, that you can change the software or use pieces of
|
||||||
|
it in new free programs, and that you know you can do these things.
|
||||||
|
|
||||||
|
To protect your rights, we need to prevent others from denying you these rights or
|
||||||
|
asking you to surrender the rights. Therefore, you have certain responsibilities if
|
||||||
|
you distribute copies of the software, or if you modify it: responsibilities to
|
||||||
|
respect the freedom of others.
|
||||||
|
|
||||||
|
For example, if you distribute copies of such a program, whether gratis or for a fee,
|
||||||
|
you must pass on to the recipients the same freedoms that you received. You must make
|
||||||
|
sure that they, too, receive or can get the source code. And you must show them these
|
||||||
|
terms so they know their rights.
|
||||||
|
|
||||||
|
Developers that use the GNU GPL protect your rights with two steps: **(1)** assert
|
||||||
|
copyright on the software, and **(2)** offer you this License giving you legal permission
|
||||||
|
to copy, distribute and/or modify it.
|
||||||
|
|
||||||
|
For the developers' and authors' protection, the GPL clearly explains that there is
|
||||||
|
no warranty for this free software. For both users' and authors' sake, the GPL
|
||||||
|
requires that modified versions be marked as changed, so that their problems will not
|
||||||
|
be attributed erroneously to authors of previous versions.
|
||||||
|
|
||||||
|
Some devices are designed to deny users access to install or run modified versions of
|
||||||
|
the software inside them, although the manufacturer can do so. This is fundamentally
|
||||||
|
incompatible with the aim of protecting users' freedom to change the software. The
|
||||||
|
systematic pattern of such abuse occurs in the area of products for individuals to
|
||||||
|
use, which is precisely where it is most unacceptable. Therefore, we have designed
|
||||||
|
this version of the GPL to prohibit the practice for those products. If such problems
|
||||||
|
arise substantially in other domains, we stand ready to extend this provision to
|
||||||
|
those domains in future versions of the GPL, as needed to protect the freedom of
|
||||||
|
users.
|
||||||
|
|
||||||
|
Finally, every program is threatened constantly by software patents. States should
|
||||||
|
not allow patents to restrict development and use of software on general-purpose
|
||||||
|
computers, but in those that do, we wish to avoid the special danger that patents
|
||||||
|
applied to a free program could make it effectively proprietary. To prevent this, the
|
||||||
|
GPL assures that patents cannot be used to render the program non-free.
|
||||||
|
|
||||||
|
The precise terms and conditions for copying, distribution and modification follow.
|
||||||
|
|
||||||
|
## TERMS AND CONDITIONS
|
||||||
|
|
||||||
|
### 0. Definitions
|
||||||
|
|
||||||
|
“This License” refers to version 3 of the GNU General Public License.
|
||||||
|
|
||||||
|
“Copyright” also means copyright-like laws that apply to other kinds of
|
||||||
|
works, such as semiconductor masks.
|
||||||
|
|
||||||
|
“The Program” refers to any copyrightable work licensed under this
|
||||||
|
License. Each licensee is addressed as “you”. “Licensees” and
|
||||||
|
“recipients” may be individuals or organizations.
|
||||||
|
|
||||||
|
To “modify” a work means to copy from or adapt all or part of the work in
|
||||||
|
a fashion requiring copyright permission, other than the making of an exact copy. The
|
||||||
|
resulting work is called a “modified version” of the earlier work or a
|
||||||
|
work “based on” the earlier work.
|
||||||
|
|
||||||
|
A “covered work” means either the unmodified Program or a work based on
|
||||||
|
the Program.
|
||||||
|
|
||||||
|
To “propagate” a work means to do anything with it that, without
|
||||||
|
permission, would make you directly or secondarily liable for infringement under
|
||||||
|
applicable copyright law, except executing it on a computer or modifying a private
|
||||||
|
copy. Propagation includes copying, distribution (with or without modification),
|
||||||
|
making available to the public, and in some countries other activities as well.
|
||||||
|
|
||||||
|
To “convey” a work means any kind of propagation that enables other
|
||||||
|
parties to make or receive copies. Mere interaction with a user through a computer
|
||||||
|
network, with no transfer of a copy, is not conveying.
|
||||||
|
|
||||||
|
An interactive user interface displays “Appropriate Legal Notices” to the
|
||||||
|
extent that it includes a convenient and prominently visible feature that **(1)**
|
||||||
|
displays an appropriate copyright notice, and **(2)** tells the user that there is no
|
||||||
|
warranty for the work (except to the extent that warranties are provided), that
|
||||||
|
licensees may convey the work under this License, and how to view a copy of this
|
||||||
|
License. If the interface presents a list of user commands or options, such as a
|
||||||
|
menu, a prominent item in the list meets this criterion.
|
||||||
|
|
||||||
|
### 1. Source Code
|
||||||
|
|
||||||
|
The “source code” for a work means the preferred form of the work for
|
||||||
|
making modifications to it. “Object code” means any non-source form of a
|
||||||
|
work.
|
||||||
|
|
||||||
|
A “Standard Interface” means an interface that either is an official
|
||||||
|
standard defined by a recognized standards body, or, in the case of interfaces
|
||||||
|
specified for a particular programming language, one that is widely used among
|
||||||
|
developers working in that language.
|
||||||
|
|
||||||
|
The “System Libraries” of an executable work include anything, other than
|
||||||
|
the work as a whole, that **(a)** is included in the normal form of packaging a Major
|
||||||
|
Component, but which is not part of that Major Component, and **(b)** serves only to
|
||||||
|
enable use of the work with that Major Component, or to implement a Standard
|
||||||
|
Interface for which an implementation is available to the public in source code form.
|
||||||
|
A “Major Component”, in this context, means a major essential component
|
||||||
|
(kernel, window system, and so on) of the specific operating system (if any) on which
|
||||||
|
the executable work runs, or a compiler used to produce the work, or an object code
|
||||||
|
interpreter used to run it.
|
||||||
|
|
||||||
|
The “Corresponding Source” for a work in object code form means all the
|
||||||
|
source code needed to generate, install, and (for an executable work) run the object
|
||||||
|
code and to modify the work, including scripts to control those activities. However,
|
||||||
|
it does not include the work's System Libraries, or general-purpose tools or
|
||||||
|
generally available free programs which are used unmodified in performing those
|
||||||
|
activities but which are not part of the work. For example, Corresponding Source
|
||||||
|
includes interface definition files associated with source files for the work, and
|
||||||
|
the source code for shared libraries and dynamically linked subprograms that the work
|
||||||
|
is specifically designed to require, such as by intimate data communication or
|
||||||
|
control flow between those subprograms and other parts of the work.
|
||||||
|
|
||||||
|
The Corresponding Source need not include anything that users can regenerate
|
||||||
|
automatically from other parts of the Corresponding Source.
|
||||||
|
|
||||||
|
The Corresponding Source for a work in source code form is that same work.
|
||||||
|
|
||||||
|
### 2. Basic Permissions
|
||||||
|
|
||||||
|
All rights granted under this License are granted for the term of copyright on the
|
||||||
|
Program, and are irrevocable provided the stated conditions are met. This License
|
||||||
|
explicitly affirms your unlimited permission to run the unmodified Program. The
|
||||||
|
output from running a covered work is covered by this License only if the output,
|
||||||
|
given its content, constitutes a covered work. This License acknowledges your rights
|
||||||
|
of fair use or other equivalent, as provided by copyright law.
|
||||||
|
|
||||||
|
You may make, run and propagate covered works that you do not convey, without
|
||||||
|
conditions so long as your license otherwise remains in force. You may convey covered
|
||||||
|
works to others for the sole purpose of having them make modifications exclusively
|
||||||
|
for you, or provide you with facilities for running those works, provided that you
|
||||||
|
comply with the terms of this License in conveying all material for which you do not
|
||||||
|
control copyright. Those thus making or running the covered works for you must do so
|
||||||
|
exclusively on your behalf, under your direction and control, on terms that prohibit
|
||||||
|
them from making any copies of your copyrighted material outside their relationship
|
||||||
|
with you.
|
||||||
|
|
||||||
|
Conveying under any other circumstances is permitted solely under the conditions
|
||||||
|
stated below. Sublicensing is not allowed; section 10 makes it unnecessary.
|
||||||
|
|
||||||
|
### 3. Protecting Users' Legal Rights From Anti-Circumvention Law
|
||||||
|
|
||||||
|
No covered work shall be deemed part of an effective technological measure under any
|
||||||
|
applicable law fulfilling obligations under article 11 of the WIPO copyright treaty
|
||||||
|
adopted on 20 December 1996, or similar laws prohibiting or restricting circumvention
|
||||||
|
of such measures.
|
||||||
|
|
||||||
|
When you convey a covered work, you waive any legal power to forbid circumvention of
|
||||||
|
technological measures to the extent such circumvention is effected by exercising
|
||||||
|
rights under this License with respect to the covered work, and you disclaim any
|
||||||
|
intention to limit operation or modification of the work as a means of enforcing,
|
||||||
|
against the work's users, your or third parties' legal rights to forbid circumvention
|
||||||
|
of technological measures.
|
||||||
|
|
||||||
|
### 4. Conveying Verbatim Copies
|
||||||
|
|
||||||
|
You may convey verbatim copies of the Program's source code as you receive it, in any
|
||||||
|
medium, provided that you conspicuously and appropriately publish on each copy an
|
||||||
|
appropriate copyright notice; keep intact all notices stating that this License and
|
||||||
|
any non-permissive terms added in accord with section 7 apply to the code; keep
|
||||||
|
intact all notices of the absence of any warranty; and give all recipients a copy of
|
||||||
|
this License along with the Program.
|
||||||
|
|
||||||
|
You may charge any price or no price for each copy that you convey, and you may offer
|
||||||
|
support or warranty protection for a fee.
|
||||||
|
|
||||||
|
### 5. Conveying Modified Source Versions
|
||||||
|
|
||||||
|
You may convey a work based on the Program, or the modifications to produce it from
|
||||||
|
the Program, in the form of source code under the terms of section 4, provided that
|
||||||
|
you also meet all of these conditions:
|
||||||
|
|
||||||
|
* **a)** The work must carry prominent notices stating that you modified it, and giving a
|
||||||
|
relevant date.
|
||||||
|
* **b)** The work must carry prominent notices stating that it is released under this
|
||||||
|
License and any conditions added under section 7. This requirement modifies the
|
||||||
|
requirement in section 4 to “keep intact all notices”.
|
||||||
|
* **c)** You must license the entire work, as a whole, under this License to anyone who
|
||||||
|
comes into possession of a copy. This License will therefore apply, along with any
|
||||||
|
applicable section 7 additional terms, to the whole of the work, and all its parts,
|
||||||
|
regardless of how they are packaged. This License gives no permission to license the
|
||||||
|
work in any other way, but it does not invalidate such permission if you have
|
||||||
|
separately received it.
|
||||||
|
* **d)** If the work has interactive user interfaces, each must display Appropriate Legal
|
||||||
|
Notices; however, if the Program has interactive interfaces that do not display
|
||||||
|
Appropriate Legal Notices, your work need not make them do so.
|
||||||
|
|
||||||
|
A compilation of a covered work with other separate and independent works, which are
|
||||||
|
not by their nature extensions of the covered work, and which are not combined with
|
||||||
|
it such as to form a larger program, in or on a volume of a storage or distribution
|
||||||
|
medium, is called an “aggregate” if the compilation and its resulting
|
||||||
|
copyright are not used to limit the access or legal rights of the compilation's users
|
||||||
|
beyond what the individual works permit. Inclusion of a covered work in an aggregate
|
||||||
|
does not cause this License to apply to the other parts of the aggregate.
|
||||||
|
|
||||||
|
### 6. Conveying Non-Source Forms
|
||||||
|
|
||||||
|
You may convey a covered work in object code form under the terms of sections 4 and
|
||||||
|
5, provided that you also convey the machine-readable Corresponding Source under the
|
||||||
|
terms of this License, in one of these ways:
|
||||||
|
|
||||||
|
* **a)** Convey the object code in, or embodied in, a physical product (including a
|
||||||
|
physical distribution medium), accompanied by the Corresponding Source fixed on a
|
||||||
|
durable physical medium customarily used for software interchange.
|
||||||
|
* **b)** Convey the object code in, or embodied in, a physical product (including a
|
||||||
|
physical distribution medium), accompanied by a written offer, valid for at least
|
||||||
|
three years and valid for as long as you offer spare parts or customer support for
|
||||||
|
that product model, to give anyone who possesses the object code either **(1)** a copy of
|
||||||
|
the Corresponding Source for all the software in the product that is covered by this
|
||||||
|
License, on a durable physical medium customarily used for software interchange, for
|
||||||
|
a price no more than your reasonable cost of physically performing this conveying of
|
||||||
|
source, or **(2)** access to copy the Corresponding Source from a network server at no
|
||||||
|
charge.
|
||||||
|
* **c)** Convey individual copies of the object code with a copy of the written offer to
|
||||||
|
provide the Corresponding Source. This alternative is allowed only occasionally and
|
||||||
|
noncommercially, and only if you received the object code with such an offer, in
|
||||||
|
accord with subsection 6b.
|
||||||
|
* **d)** Convey the object code by offering access from a designated place (gratis or for
|
||||||
|
a charge), and offer equivalent access to the Corresponding Source in the same way
|
||||||
|
through the same place at no further charge. You need not require recipients to copy
|
||||||
|
the Corresponding Source along with the object code. If the place to copy the object
|
||||||
|
code is a network server, the Corresponding Source may be on a different server
|
||||||
|
(operated by you or a third party) that supports equivalent copying facilities,
|
||||||
|
provided you maintain clear directions next to the object code saying where to find
|
||||||
|
the Corresponding Source. Regardless of what server hosts the Corresponding Source,
|
||||||
|
you remain obligated to ensure that it is available for as long as needed to satisfy
|
||||||
|
these requirements.
|
||||||
|
* **e)** Convey the object code using peer-to-peer transmission, provided you inform
|
||||||
|
other peers where the object code and Corresponding Source of the work are being
|
||||||
|
offered to the general public at no charge under subsection 6d.
|
||||||
|
|
||||||
|
A separable portion of the object code, whose source code is excluded from the
|
||||||
|
Corresponding Source as a System Library, need not be included in conveying the
|
||||||
|
object code work.
|
||||||
|
|
||||||
|
A “User Product” is either **(1)** a “consumer product”, which
|
||||||
|
means any tangible personal property which is normally used for personal, family, or
|
||||||
|
household purposes, or **(2)** anything designed or sold for incorporation into a
|
||||||
|
dwelling. In determining whether a product is a consumer product, doubtful cases
|
||||||
|
shall be resolved in favor of coverage. For a particular product received by a
|
||||||
|
particular user, “normally used” refers to a typical or common use of
|
||||||
|
that class of product, regardless of the status of the particular user or of the way
|
||||||
|
in which the particular user actually uses, or expects or is expected to use, the
|
||||||
|
product. A product is a consumer product regardless of whether the product has
|
||||||
|
substantial commercial, industrial or non-consumer uses, unless such uses represent
|
||||||
|
the only significant mode of use of the product.
|
||||||
|
|
||||||
|
“Installation Information” for a User Product means any methods,
|
||||||
|
procedures, authorization keys, or other information required to install and execute
|
||||||
|
modified versions of a covered work in that User Product from a modified version of
|
||||||
|
its Corresponding Source. The information must suffice to ensure that the continued
|
||||||
|
functioning of the modified object code is in no case prevented or interfered with
|
||||||
|
solely because modification has been made.
|
||||||
|
|
||||||
|
If you convey an object code work under this section in, or with, or specifically for
|
||||||
|
use in, a User Product, and the conveying occurs as part of a transaction in which
|
||||||
|
the right of possession and use of the User Product is transferred to the recipient
|
||||||
|
in perpetuity or for a fixed term (regardless of how the transaction is
|
||||||
|
characterized), the Corresponding Source conveyed under this section must be
|
||||||
|
accompanied by the Installation Information. But this requirement does not apply if
|
||||||
|
neither you nor any third party retains the ability to install modified object code
|
||||||
|
on the User Product (for example, the work has been installed in ROM).
|
||||||
|
|
||||||
|
The requirement to provide Installation Information does not include a requirement to
|
||||||
|
continue to provide support service, warranty, or updates for a work that has been
|
||||||
|
modified or installed by the recipient, or for the User Product in which it has been
|
||||||
|
modified or installed. Access to a network may be denied when the modification itself
|
||||||
|
materially and adversely affects the operation of the network or violates the rules
|
||||||
|
and protocols for communication across the network.
|
||||||
|
|
||||||
|
Corresponding Source conveyed, and Installation Information provided, in accord with
|
||||||
|
this section must be in a format that is publicly documented (and with an
|
||||||
|
implementation available to the public in source code form), and must require no
|
||||||
|
special password or key for unpacking, reading or copying.
|
||||||
|
|
||||||
|
### 7. Additional Terms
|
||||||
|
|
||||||
|
“Additional permissions” are terms that supplement the terms of this
|
||||||
|
License by making exceptions from one or more of its conditions. Additional
|
||||||
|
permissions that are applicable to the entire Program shall be treated as though they
|
||||||
|
were included in this License, to the extent that they are valid under applicable
|
||||||
|
law. If additional permissions apply only to part of the Program, that part may be
|
||||||
|
used separately under those permissions, but the entire Program remains governed by
|
||||||
|
this License without regard to the additional permissions.
|
||||||
|
|
||||||
|
When you convey a copy of a covered work, you may at your option remove any
|
||||||
|
additional permissions from that copy, or from any part of it. (Additional
|
||||||
|
permissions may be written to require their own removal in certain cases when you
|
||||||
|
modify the work.) You may place additional permissions on material, added by you to a
|
||||||
|
covered work, for which you have or can give appropriate copyright permission.
|
||||||
|
|
||||||
|
Notwithstanding any other provision of this License, for material you add to a
|
||||||
|
covered work, you may (if authorized by the copyright holders of that material)
|
||||||
|
supplement the terms of this License with terms:
|
||||||
|
|
||||||
|
* **a)** Disclaiming warranty or limiting liability differently from the terms of
|
||||||
|
sections 15 and 16 of this License; or
|
||||||
|
* **b)** Requiring preservation of specified reasonable legal notices or author
|
||||||
|
attributions in that material or in the Appropriate Legal Notices displayed by works
|
||||||
|
containing it; or
|
||||||
|
* **c)** Prohibiting misrepresentation of the origin of that material, or requiring that
|
||||||
|
modified versions of such material be marked in reasonable ways as different from the
|
||||||
|
original version; or
|
||||||
|
* **d)** Limiting the use for publicity purposes of names of licensors or authors of the
|
||||||
|
material; or
|
||||||
|
* **e)** Declining to grant rights under trademark law for use of some trade names,
|
||||||
|
trademarks, or service marks; or
|
||||||
|
* **f)** Requiring indemnification of licensors and authors of that material by anyone
|
||||||
|
who conveys the material (or modified versions of it) with contractual assumptions of
|
||||||
|
liability to the recipient, for any liability that these contractual assumptions
|
||||||
|
directly impose on those licensors and authors.
|
||||||
|
|
||||||
|
All other non-permissive additional terms are considered “further
|
||||||
|
restrictions” within the meaning of section 10. If the Program as you received
|
||||||
|
it, or any part of it, contains a notice stating that it is governed by this License
|
||||||
|
along with a term that is a further restriction, you may remove that term. If a
|
||||||
|
license document contains a further restriction but permits relicensing or conveying
|
||||||
|
under this License, you may add to a covered work material governed by the terms of
|
||||||
|
that license document, provided that the further restriction does not survive such
|
||||||
|
relicensing or conveying.
|
||||||
|
|
||||||
|
If you add terms to a covered work in accord with this section, you must place, in
|
||||||
|
the relevant source files, a statement of the additional terms that apply to those
|
||||||
|
files, or a notice indicating where to find the applicable terms.
|
||||||
|
|
||||||
|
Additional terms, permissive or non-permissive, may be stated in the form of a
|
||||||
|
separately written license, or stated as exceptions; the above requirements apply
|
||||||
|
either way.
|
||||||
|
|
||||||
|
### 8. Termination
|
||||||
|
|
||||||
|
You may not propagate or modify a covered work except as expressly provided under
|
||||||
|
this License. Any attempt otherwise to propagate or modify it is void, and will
|
||||||
|
automatically terminate your rights under this License (including any patent licenses
|
||||||
|
granted under the third paragraph of section 11).
|
||||||
|
|
||||||
|
However, if you cease all violation of this License, then your license from a
|
||||||
|
particular copyright holder is reinstated **(a)** provisionally, unless and until the
|
||||||
|
copyright holder explicitly and finally terminates your license, and **(b)** permanently,
|
||||||
|
if the copyright holder fails to notify you of the violation by some reasonable means
|
||||||
|
prior to 60 days after the cessation.
|
||||||
|
|
||||||
|
Moreover, your license from a particular copyright holder is reinstated permanently
|
||||||
|
if the copyright holder notifies you of the violation by some reasonable means, this
|
||||||
|
is the first time you have received notice of violation of this License (for any
|
||||||
|
work) from that copyright holder, and you cure the violation prior to 30 days after
|
||||||
|
your receipt of the notice.
|
||||||
|
|
||||||
|
Termination of your rights under this section does not terminate the licenses of
|
||||||
|
parties who have received copies or rights from you under this License. If your
|
||||||
|
rights have been terminated and not permanently reinstated, you do not qualify to
|
||||||
|
receive new licenses for the same material under section 10.
|
||||||
|
|
||||||
|
### 9. Acceptance Not Required for Having Copies
|
||||||
|
|
||||||
|
You are not required to accept this License in order to receive or run a copy of the
|
||||||
|
Program. Ancillary propagation of a covered work occurring solely as a consequence of
|
||||||
|
using peer-to-peer transmission to receive a copy likewise does not require
|
||||||
|
acceptance. However, nothing other than this License grants you permission to
|
||||||
|
propagate or modify any covered work. These actions infringe copyright if you do not
|
||||||
|
accept this License. Therefore, by modifying or propagating a covered work, you
|
||||||
|
indicate your acceptance of this License to do so.
|
||||||
|
|
||||||
|
### 10. Automatic Licensing of Downstream Recipients
|
||||||
|
|
||||||
|
Each time you convey a covered work, the recipient automatically receives a license
|
||||||
|
from the original licensors, to run, modify and propagate that work, subject to this
|
||||||
|
License. You are not responsible for enforcing compliance by third parties with this
|
||||||
|
License.
|
||||||
|
|
||||||
|
An “entity transaction” is a transaction transferring control of an
|
||||||
|
organization, or substantially all assets of one, or subdividing an organization, or
|
||||||
|
merging organizations. If propagation of a covered work results from an entity
|
||||||
|
transaction, each party to that transaction who receives a copy of the work also
|
||||||
|
receives whatever licenses to the work the party's predecessor in interest had or
|
||||||
|
could give under the previous paragraph, plus a right to possession of the
|
||||||
|
Corresponding Source of the work from the predecessor in interest, if the predecessor
|
||||||
|
has it or can get it with reasonable efforts.
|
||||||
|
|
||||||
|
You may not impose any further restrictions on the exercise of the rights granted or
|
||||||
|
affirmed under this License. For example, you may not impose a license fee, royalty,
|
||||||
|
or other charge for exercise of rights granted under this License, and you may not
|
||||||
|
initiate litigation (including a cross-claim or counterclaim in a lawsuit) alleging
|
||||||
|
that any patent claim is infringed by making, using, selling, offering for sale, or
|
||||||
|
importing the Program or any portion of it.
|
||||||
|
|
||||||
|
### 11. Patents
|
||||||
|
|
||||||
|
A “contributor” is a copyright holder who authorizes use under this
|
||||||
|
License of the Program or a work on which the Program is based. The work thus
|
||||||
|
licensed is called the contributor's “contributor version”.
|
||||||
|
|
||||||
|
A contributor's “essential patent claims” are all patent claims owned or
|
||||||
|
controlled by the contributor, whether already acquired or hereafter acquired, that
|
||||||
|
would be infringed by some manner, permitted by this License, of making, using, or
|
||||||
|
selling its contributor version, but do not include claims that would be infringed
|
||||||
|
only as a consequence of further modification of the contributor version. For
|
||||||
|
purposes of this definition, “control” includes the right to grant patent
|
||||||
|
sublicenses in a manner consistent with the requirements of this License.
|
||||||
|
|
||||||
|
Each contributor grants you a non-exclusive, worldwide, royalty-free patent license
|
||||||
|
under the contributor's essential patent claims, to make, use, sell, offer for sale,
|
||||||
|
import and otherwise run, modify and propagate the contents of its contributor
|
||||||
|
version.
|
||||||
|
|
||||||
|
In the following three paragraphs, a “patent license” is any express
|
||||||
|
agreement or commitment, however denominated, not to enforce a patent (such as an
|
||||||
|
express permission to practice a patent or covenant not to sue for patent
|
||||||
|
infringement). To “grant” such a patent license to a party means to make
|
||||||
|
such an agreement or commitment not to enforce a patent against the party.
|
||||||
|
|
||||||
|
If you convey a covered work, knowingly relying on a patent license, and the
|
||||||
|
Corresponding Source of the work is not available for anyone to copy, free of charge
|
||||||
|
and under the terms of this License, through a publicly available network server or
|
||||||
|
other readily accessible means, then you must either **(1)** cause the Corresponding
|
||||||
|
Source to be so available, or **(2)** arrange to deprive yourself of the benefit of the
|
||||||
|
patent license for this particular work, or **(3)** arrange, in a manner consistent with
|
||||||
|
the requirements of this License, to extend the patent license to downstream
|
||||||
|
recipients. “Knowingly relying” means you have actual knowledge that, but
|
||||||
|
for the patent license, your conveying the covered work in a country, or your
|
||||||
|
recipient's use of the covered work in a country, would infringe one or more
|
||||||
|
identifiable patents in that country that you have reason to believe are valid.
|
||||||
|
|
||||||
|
If, pursuant to or in connection with a single transaction or arrangement, you
|
||||||
|
convey, or propagate by procuring conveyance of, a covered work, and grant a patent
|
||||||
|
license to some of the parties receiving the covered work authorizing them to use,
|
||||||
|
propagate, modify or convey a specific copy of the covered work, then the patent
|
||||||
|
license you grant is automatically extended to all recipients of the covered work and
|
||||||
|
works based on it.
|
||||||
|
|
||||||
|
A patent license is “discriminatory” if it does not include within the
|
||||||
|
scope of its coverage, prohibits the exercise of, or is conditioned on the
|
||||||
|
non-exercise of one or more of the rights that are specifically granted under this
|
||||||
|
License. You may not convey a covered work if you are a party to an arrangement with
|
||||||
|
a third party that is in the business of distributing software, under which you make
|
||||||
|
payment to the third party based on the extent of your activity of conveying the
|
||||||
|
work, and under which the third party grants, to any of the parties who would receive
|
||||||
|
the covered work from you, a discriminatory patent license **(a)** in connection with
|
||||||
|
copies of the covered work conveyed by you (or copies made from those copies), or **(b)**
|
||||||
|
primarily for and in connection with specific products or compilations that contain
|
||||||
|
the covered work, unless you entered into that arrangement, or that patent license
|
||||||
|
was granted, prior to 28 March 2007.
|
||||||
|
|
||||||
|
Nothing in this License shall be construed as excluding or limiting any implied
|
||||||
|
license or other defenses to infringement that may otherwise be available to you
|
||||||
|
under applicable patent law.
|
||||||
|
|
||||||
|
### 12. No Surrender of Others' Freedom
|
||||||
|
|
||||||
|
If conditions are imposed on you (whether by court order, agreement or otherwise)
|
||||||
|
that contradict the conditions of this License, they do not excuse you from the
|
||||||
|
conditions of this License. If you cannot convey a covered work so as to satisfy
|
||||||
|
simultaneously your obligations under this License and any other pertinent
|
||||||
|
obligations, then as a consequence you may not convey it at all. For example, if you
|
||||||
|
agree to terms that obligate you to collect a royalty for further conveying from
|
||||||
|
those to whom you convey the Program, the only way you could satisfy both those terms
|
||||||
|
and this License would be to refrain entirely from conveying the Program.
|
||||||
|
|
||||||
|
### 13. Use with the GNU Affero General Public License
|
||||||
|
|
||||||
|
Notwithstanding any other provision of this License, you have permission to link or
|
||||||
|
combine any covered work with a work licensed under version 3 of the GNU Affero
|
||||||
|
General Public License into a single combined work, and to convey the resulting work.
|
||||||
|
The terms of this License will continue to apply to the part which is the covered
|
||||||
|
work, but the special requirements of the GNU Affero General Public License, section
|
||||||
|
13, concerning interaction through a network will apply to the combination as such.
|
||||||
|
|
||||||
|
### 14. Revised Versions of this License
|
||||||
|
|
||||||
|
The Free Software Foundation may publish revised and/or new versions of the GNU
|
||||||
|
General Public License from time to time. Such new versions will be similar in spirit
|
||||||
|
to the present version, but may differ in detail to address new problems or concerns.
|
||||||
|
|
||||||
|
Each version is given a distinguishing version number. If the Program specifies that
|
||||||
|
a certain numbered version of the GNU General Public License “or any later
|
||||||
|
version” applies to it, you have the option of following the terms and
|
||||||
|
conditions either of that numbered version or of any later version published by the
|
||||||
|
Free Software Foundation. If the Program does not specify a version number of the GNU
|
||||||
|
General Public License, you may choose any version ever published by the Free
|
||||||
|
Software Foundation.
|
||||||
|
|
||||||
|
If the Program specifies that a proxy can decide which future versions of the GNU
|
||||||
|
General Public License can be used, that proxy's public statement of acceptance of a
|
||||||
|
version permanently authorizes you to choose that version for the Program.
|
||||||
|
|
||||||
|
Later license versions may give you additional or different permissions. However, no
|
||||||
|
additional obligations are imposed on any author or copyright holder as a result of
|
||||||
|
your choosing to follow a later version.
|
||||||
|
|
||||||
|
### 15. Disclaimer of Warranty
|
||||||
|
|
||||||
|
THERE IS NO WARRANTY FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW.
|
||||||
|
EXCEPT WHEN OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES
|
||||||
|
PROVIDE THE PROGRAM “AS IS” WITHOUT WARRANTY OF ANY KIND, EITHER
|
||||||
|
EXPRESSED OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
|
||||||
|
MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS TO THE
|
||||||
|
QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE PROGRAM PROVE
|
||||||
|
DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, REPAIR OR CORRECTION.
|
||||||
|
|
||||||
|
### 16. Limitation of Liability
|
||||||
|
|
||||||
|
IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING WILL ANY
|
||||||
|
COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MODIFIES AND/OR CONVEYS THE PROGRAM AS
|
||||||
|
PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, INCLUDING ANY GENERAL, SPECIAL,
|
||||||
|
INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING OUT OF THE USE OR INABILITY TO USE THE
|
||||||
|
PROGRAM (INCLUDING BUT NOT LIMITED TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE
|
||||||
|
OR LOSSES SUSTAINED BY YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE
|
||||||
|
WITH ANY OTHER PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE
|
||||||
|
POSSIBILITY OF SUCH DAMAGES.
|
||||||
|
|
||||||
|
### 17. Interpretation of Sections 15 and 16
|
||||||
|
|
||||||
|
If the disclaimer of warranty and limitation of liability provided above cannot be
|
||||||
|
given local legal effect according to their terms, reviewing courts shall apply local
|
||||||
|
law that most closely approximates an absolute waiver of all civil liability in
|
||||||
|
connection with the Program, unless a warranty or assumption of liability accompanies
|
||||||
|
a copy of the Program in return for a fee.
|
||||||
|
|
||||||
|
_END OF TERMS AND CONDITIONS_
|
||||||
|
|
||||||
|
## How to Apply These Terms to Your New Programs
|
||||||
|
|
||||||
|
If you develop a new program, and you want it to be of the greatest possible use to
|
||||||
|
the public, the best way to achieve this is to make it free software which everyone
|
||||||
|
can redistribute and change under these terms.
|
||||||
|
|
||||||
|
To do so, attach the following notices to the program. It is safest to attach them
|
||||||
|
to the start of each source file to most effectively state the exclusion of warranty;
|
||||||
|
and each file should have at least the “copyright” line and a pointer to
|
||||||
|
where the full notice is found.
|
||||||
|
|
||||||
|
<one line to give the program's name and a brief idea of what it does.>
|
||||||
|
Copyright (C) <year> <name of author>
|
||||||
|
|
||||||
|
This program is free software: you can redistribute it and/or modify
|
||||||
|
it under the terms of the GNU General Public License as published by
|
||||||
|
the Free Software Foundation, either version 3 of the License, or
|
||||||
|
(at your option) any later version.
|
||||||
|
|
||||||
|
This program is distributed in the hope that it will be useful,
|
||||||
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
GNU General Public License for more details.
|
||||||
|
|
||||||
|
You should have received a copy of the GNU General Public License
|
||||||
|
along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
Also add information on how to contact you by electronic and paper mail.
|
||||||
|
|
||||||
|
If the program does terminal interaction, make it output a short notice like this
|
||||||
|
when it starts in an interactive mode:
|
||||||
|
|
||||||
|
<program> Copyright (C) <year> <name of author>
|
||||||
|
This program comes with ABSOLUTELY NO WARRANTY; for details type 'show w'.
|
||||||
|
This is free software, and you are welcome to redistribute it
|
||||||
|
under certain conditions; type 'show c' for details.
|
||||||
|
|
||||||
|
The hypothetical commands `show w` and `show c` should show the appropriate parts of
|
||||||
|
the General Public License. Of course, your program's commands might be different;
|
||||||
|
for a GUI interface, you would use an “about box”.
|
||||||
|
|
||||||
|
You should also get your employer (if you work as a programmer) or school, if any, to
|
||||||
|
sign a “copyright disclaimer” for the program, if necessary. For more
|
||||||
|
information on this, and how to apply and follow the GNU GPL, see
|
||||||
|
<<http://www.gnu.org/licenses/>>.
|
||||||
|
|
||||||
|
The GNU General Public License does not permit incorporating your program into
|
||||||
|
proprietary programs. If your program is a subroutine library, you may consider it
|
||||||
|
more useful to permit linking proprietary applications with the library. If this is
|
||||||
|
what you want to do, use the GNU Lesser General Public License instead of this
|
||||||
|
License. But first, please read
|
||||||
|
<<http://www.gnu.org/philosophy/why-not-lgpl.html>>.
|
||||||
48
README.md
Normal file
48
README.md
Normal file
@@ -0,0 +1,48 @@
|
|||||||
|
# FFX
|
||||||
|
|
||||||
|
## Installation
|
||||||
|
|
||||||
|
per https:
|
||||||
|
|
||||||
|
```sh
|
||||||
|
pip install https://<URL>/<Releaser>/ffx.git@<Branch>
|
||||||
|
```
|
||||||
|
|
||||||
|
per git:
|
||||||
|
|
||||||
|
```sh
|
||||||
|
pip install git+ssh://<Username>@<URL>/<Releaser>/ffx.git@<Branch>
|
||||||
|
```
|
||||||
|
|
||||||
|
## Version history
|
||||||
|
|
||||||
|
### 0.1.1
|
||||||
|
|
||||||
|
Bugfixes, TMBD identify shows
|
||||||
|
|
||||||
|
### 0.1.2
|
||||||
|
|
||||||
|
Bugfixes
|
||||||
|
|
||||||
|
### 0.1.3
|
||||||
|
|
||||||
|
Subtitle file imports
|
||||||
|
|
||||||
|
### 0.2.0
|
||||||
|
|
||||||
|
Tests, Config-File
|
||||||
|
|
||||||
|
### 0.2.1
|
||||||
|
|
||||||
|
Signature, Tags cleaning, Bugfixes, Refactoring
|
||||||
|
|
||||||
|
### 0.2.2
|
||||||
|
|
||||||
|
CLI-Overrides
|
||||||
|
|
||||||
|
### 0.2.3
|
||||||
|
|
||||||
|
PyPi packaging
|
||||||
|
Templating output filename
|
||||||
|
Season shiftung
|
||||||
|
DB-Versionierung
|
||||||
237
bin/ffx.py
237
bin/ffx.py
@@ -1,237 +0,0 @@
|
|||||||
#! /usr/bin/python3
|
|
||||||
|
|
||||||
import os, sys, subprocess, json
|
|
||||||
|
|
||||||
DEFAULT_QUALITY = 23
|
|
||||||
DEFAULT_AV1_PRESET = 5
|
|
||||||
DEFAULT_STEREO_BANDWIDTH = "128"
|
|
||||||
DEFAULT_AC3_BANDWIDTH = "256"
|
|
||||||
DEFAULT_DTS_BANDWIDTH = "320"
|
|
||||||
|
|
||||||
TEMP_FILE_NAME = "ffmpeg2pass-0.log"
|
|
||||||
|
|
||||||
def executeProcess(commandSequence):
|
|
||||||
|
|
||||||
process = subprocess.Popen(commandSequence, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
||||||
|
|
||||||
output, error = process.communicate()
|
|
||||||
|
|
||||||
return output
|
|
||||||
|
|
||||||
|
|
||||||
def generateAV1Tokens(q, p):
|
|
||||||
|
|
||||||
return ['-c:v:0',
|
|
||||||
'libsvtav1',
|
|
||||||
'-svtav1-params',
|
|
||||||
f"crf={q}:preset={p}:tune=0:enable-overlays=1:scd=1:scm=0",
|
|
||||||
'-pix_fmt', 'yuv420p10le']
|
|
||||||
|
|
||||||
def generateVP9Pass1Tokens(q):
|
|
||||||
|
|
||||||
return ['-c:v:0', 'libvpx-vp9', '-row-mt', '1',
|
|
||||||
'-crf', str(q), '-pass', '1', '-speed', '4',
|
|
||||||
'-frame-parallel', '0', '-g', '9999', '-aq-mode', '0']
|
|
||||||
|
|
||||||
def generateVP9Pass2Tokens(q):
|
|
||||||
|
|
||||||
return ['-c:v:0', 'libvpx-vp9', '-row-mt', '1',
|
|
||||||
'-crf', str(q), '-pass', '2',
|
|
||||||
'-frame-parallel', '0', '-g', '9999', '-aq-mode', '0',
|
|
||||||
'-auto-alt-ref', '1', '-lag-in-frames', '25']
|
|
||||||
|
|
||||||
def generateCropTokens(start, length):
|
|
||||||
|
|
||||||
return ['-ss', str(start), '-t', str(length)]
|
|
||||||
|
|
||||||
def generateOutputTokens(f, q=''):
|
|
||||||
if q:
|
|
||||||
fTokens = f.split('.')
|
|
||||||
paddedFilename = '.'.join(fTokens[:-1]) + f" q{q}" + '.' + fTokens[-1]
|
|
||||||
return ['-f', 'webm', paddedFilename]
|
|
||||||
else:
|
|
||||||
return ['-f', 'webm', f]
|
|
||||||
|
|
||||||
|
|
||||||
inputFilename = sys.argv[1]
|
|
||||||
outputFilename = sys.argv[2]
|
|
||||||
|
|
||||||
targetFormat = 'vp9'
|
|
||||||
if 'av1' in sys.argv:
|
|
||||||
targetFormat = 'av1'
|
|
||||||
if 'vp9' in sys.argv:
|
|
||||||
targetFormat = 'vp9'
|
|
||||||
|
|
||||||
|
|
||||||
qualities = [str(DEFAULT_QUALITY)]
|
|
||||||
qualitiesTokens = [q for q in sys.argv if q.startswith('q=')]
|
|
||||||
if qualitiesTokens:
|
|
||||||
qualitiesString = qualitiesTokens[0].split('=')[1]
|
|
||||||
qualities = qualitiesString.split(',')
|
|
||||||
|
|
||||||
preset = DEFAULT_AV1_PRESET
|
|
||||||
presetTokens = [p for p in sys.argv if p.startswith('p=')]
|
|
||||||
if presetTokens:
|
|
||||||
preset = int(presetTokens[0].split('=')[1])
|
|
||||||
|
|
||||||
stereoBandwidth = DEFAULT_STEREO_BANDWIDTH
|
|
||||||
stereoTokens = [s for s in sys.argv if s.startswith('a=')]
|
|
||||||
if stereoTokens:
|
|
||||||
stereoBandwidth = str(stereoTokens[0].split('=')[1])
|
|
||||||
if not stereoBandwidth.endswith('k'):
|
|
||||||
stereoBandwidth += "k"
|
|
||||||
|
|
||||||
ac3Bandwidth = DEFAULT_AC3_BANDWIDTH
|
|
||||||
ac3Tokens = [a for a in sys.argv if a.startswith('ac3=')]
|
|
||||||
if ac3Tokens:
|
|
||||||
ac3Bandwidth = str(ac3Tokens[0].split('=')[1])
|
|
||||||
if not ac3Bandwidth.endswith('k'):
|
|
||||||
ac3Bandwidth += "k"
|
|
||||||
|
|
||||||
dtsBandwidth = DEFAULT_DTS_BANDWIDTH
|
|
||||||
dtsTokens = [d for d in sys.argv if d.startswith('dts=')]
|
|
||||||
if dtsTokens:
|
|
||||||
dtsBandwidth = str(dtsTokens[0].split('=')[1])
|
|
||||||
if not dtsBandwidth.endswith('k'):
|
|
||||||
dtsBandwidth += "k"
|
|
||||||
|
|
||||||
cropStart = ''
|
|
||||||
cropLength = ''
|
|
||||||
cropTokens = [c for c in sys.argv if c.startswith('crop')]
|
|
||||||
if cropTokens:
|
|
||||||
if '=' in cropTokens[0]:
|
|
||||||
cropString = cropTokens[0].split('=')[1]
|
|
||||||
cropStart, cropLength = cropString.split(',')
|
|
||||||
else:
|
|
||||||
cropStart = 60
|
|
||||||
cropLength = 180
|
|
||||||
|
|
||||||
|
|
||||||
output = executeProcess(["ffprobe", "-show_streams", "-of", "json" ,inputFilename])
|
|
||||||
|
|
||||||
streamData = json.loads(output)['streams']
|
|
||||||
|
|
||||||
videoStreams = [s for s in streamData if s['codec_type'] == 'video']
|
|
||||||
audioStreams = [s for s in streamData if s['codec_type'] == 'audio']
|
|
||||||
subtitleStreams = [s for s in streamData if s['codec_type'] == 'subtitle']
|
|
||||||
|
|
||||||
for aStream in audioStreams:
|
|
||||||
if 'channel_layout' in aStream:
|
|
||||||
print(f"audio stream: {aStream['channel_layout']}") #channel_layout
|
|
||||||
else:
|
|
||||||
print(f"unknown audio stream with {aStream['channels']} channels") #channel_layout
|
|
||||||
|
|
||||||
commandTokens = ['ffmpeg', '-y', '-i', inputFilename]
|
|
||||||
|
|
||||||
|
|
||||||
mappingTokens = ['-map', 'v:0']
|
|
||||||
|
|
||||||
for a in range(len(audioStreams)):
|
|
||||||
mappingTokens += ['-map', f"a:{a}"]
|
|
||||||
|
|
||||||
for s in range(len(subtitleStreams)):
|
|
||||||
mappingTokens += ['-map', f"s:{s}"]
|
|
||||||
|
|
||||||
|
|
||||||
audioTokens = []
|
|
||||||
audioStreamIndex = 0
|
|
||||||
for aStream in audioStreams:
|
|
||||||
|
|
||||||
channels = aStream['channels']
|
|
||||||
|
|
||||||
if 'channel_layout' in aStream.keys():
|
|
||||||
|
|
||||||
channelLayout = aStream['channel_layout']
|
|
||||||
|
|
||||||
if channelLayout == '6.1':
|
|
||||||
audioTokens += [f"-c:a:{audioStreamIndex}",
|
|
||||||
'libopus',
|
|
||||||
f"-filter:a:{audioStreamIndex}",
|
|
||||||
'channelmap=channel_layout=6.1',
|
|
||||||
f"-b:a:{audioStreamIndex}",
|
|
||||||
dtsBandwidth]
|
|
||||||
|
|
||||||
if channelLayout == '5.1(side)':
|
|
||||||
audioTokens += [f"-c:a:{audioStreamIndex}",
|
|
||||||
'libopus',
|
|
||||||
f"-filter:a:{audioStreamIndex}",
|
|
||||||
"channelmap=FL-FL|FR-FR|FC-FC|LFE-LFE|SL-BL|SR-BR:5.1",
|
|
||||||
f"-b:a:{audioStreamIndex}",
|
|
||||||
ac3Bandwidth]
|
|
||||||
|
|
||||||
if channelLayout == 'stereo':
|
|
||||||
audioTokens += [f"-c:a:{audioStreamIndex}",
|
|
||||||
'libopus',
|
|
||||||
f"-b:a:{audioStreamIndex}",
|
|
||||||
stereoBandwidth]
|
|
||||||
else:
|
|
||||||
|
|
||||||
if channels == 6:
|
|
||||||
audioTokens += [f"-c:a:{audioStreamIndex}",
|
|
||||||
'libopus',
|
|
||||||
f"-filter:a:{audioStreamIndex}",
|
|
||||||
"channelmap=FL-FL|FR-FR|FC-FC|LFE-LFE|SL-BL|SR-BR:5.1",
|
|
||||||
f"-b:a:{audioStreamIndex}",
|
|
||||||
ac3Bandwidth]
|
|
||||||
|
|
||||||
|
|
||||||
audioStreamIndex += 1
|
|
||||||
|
|
||||||
|
|
||||||
nullTokens = ['-f', 'null', '/dev/null']
|
|
||||||
|
|
||||||
|
|
||||||
for quality in qualities:
|
|
||||||
|
|
||||||
if targetFormat == 'av1':
|
|
||||||
|
|
||||||
commandSequence = commandTokens + mappingTokens + generateAV1Tokens(quality, preset) + audioTokens
|
|
||||||
|
|
||||||
if cropStart:
|
|
||||||
commandSequence += generateCropTokens(cropStart, cropLength)
|
|
||||||
|
|
||||||
if len(qualities) > 1:
|
|
||||||
commandSequence += generateOutputTokens(outputFilename, quality)
|
|
||||||
else:
|
|
||||||
commandSequence += generateOutputTokens(outputFilename)
|
|
||||||
|
|
||||||
print(f"Command: {' '.join(commandSequence)}")
|
|
||||||
|
|
||||||
executeProcess(commandSequence)
|
|
||||||
|
|
||||||
|
|
||||||
if targetFormat == 'vp9':
|
|
||||||
|
|
||||||
commandSequence1 = commandTokens + mappingTokens + generateVP9Pass1Tokens(quality)
|
|
||||||
|
|
||||||
|
|
||||||
if cropStart:
|
|
||||||
commandSequence1 += generateCropTokens(cropStart, cropLength)
|
|
||||||
|
|
||||||
commandSequence1 += nullTokens
|
|
||||||
|
|
||||||
|
|
||||||
print(f"Command 1: {' '.join(commandSequence1)}")
|
|
||||||
|
|
||||||
if os.path.exists(TEMP_FILE_NAME):
|
|
||||||
os.remove(TEMP_FILE_NAME)
|
|
||||||
|
|
||||||
executeProcess(commandSequence1)
|
|
||||||
|
|
||||||
commandSequence2 = commandTokens + mappingTokens + generateVP9Pass2Tokens(quality) + audioTokens
|
|
||||||
|
|
||||||
|
|
||||||
if cropStart:
|
|
||||||
commandSequence2 += generateCropTokens(cropStart, cropLength)
|
|
||||||
|
|
||||||
|
|
||||||
if len(qualities) > 1:
|
|
||||||
commandSequence2 += generateOutputTokens(outputFilename, quality)
|
|
||||||
else:
|
|
||||||
commandSequence2 += generateOutputTokens(outputFilename)
|
|
||||||
|
|
||||||
|
|
||||||
print(f"Command 2: {' '.join(commandSequence2)}")
|
|
||||||
|
|
||||||
executeProcess(commandSequence2)
|
|
||||||
|
|
||||||
38
pyproject.toml
Normal file
38
pyproject.toml
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
[project]
|
||||||
|
name = "ffx"
|
||||||
|
description = "FFX recoding and metadata managing tool"
|
||||||
|
version = "0.2.3"
|
||||||
|
license = {file = "LICENSE.md"}
|
||||||
|
dependencies = [
|
||||||
|
"requests",
|
||||||
|
"jinja2",
|
||||||
|
"click",
|
||||||
|
"textual",
|
||||||
|
"sqlalchemy",
|
||||||
|
]
|
||||||
|
readme = {file = "README.md", content-type = "text/markdown"}
|
||||||
|
authors = [
|
||||||
|
{name = "Marius", email = "javanaut@maveno.de"}
|
||||||
|
]
|
||||||
|
maintainers = [
|
||||||
|
{name = "Marius", email = "javanaut@maveno.de"}
|
||||||
|
]
|
||||||
|
classifiers = [
|
||||||
|
"Development Status :: 3 - Alpha",
|
||||||
|
"Programming Language :: Python"
|
||||||
|
]
|
||||||
|
|
||||||
|
[project.urls]
|
||||||
|
Homepage = "https://gitea.maveno.de/Javanaut/ffx"
|
||||||
|
Repository = "https://gitea.maveno.de/Javanaut/ffx.git"
|
||||||
|
Issues = "https://gitea.maveno.de/Javanaut/ffx/issues"
|
||||||
|
|
||||||
|
[build-system]
|
||||||
|
requires = [
|
||||||
|
"setuptools",
|
||||||
|
"wheel"
|
||||||
|
]
|
||||||
|
build-backend = "setuptools.build_meta"
|
||||||
|
|
||||||
|
[project.scripts]
|
||||||
|
ffx = "ffx.ffx:ffx"
|
||||||
0
src/ffx/__init__.py
Normal file
0
src/ffx/__init__.py
Normal file
61
src/ffx/audio_layout.py
Normal file
61
src/ffx/audio_layout.py
Normal file
@@ -0,0 +1,61 @@
|
|||||||
|
from enum import Enum
|
||||||
|
from .track_type import TrackType
|
||||||
|
|
||||||
|
class AudioLayout(Enum):
|
||||||
|
|
||||||
|
LAYOUT_STEREO = {"label": "stereo", "index": 1}
|
||||||
|
LAYOUT_5_1 = {"label": "5.1(side)", "index": 2}
|
||||||
|
LAYOUT_6_1 = {"label": "6.1", "index": 3}
|
||||||
|
LAYOUT_7_1 = {"label": "7.1", "index": 4} #TODO: Does this exist?
|
||||||
|
|
||||||
|
LAYOUT_6CH = {"label": "6ch", "index": 5}
|
||||||
|
|
||||||
|
LAYOUT_UNDEFINED = {"label": "undefined", "index": 0}
|
||||||
|
|
||||||
|
|
||||||
|
def label(self):
|
||||||
|
"""Returns the audio layout as string"""
|
||||||
|
return str(self.value['label'])
|
||||||
|
|
||||||
|
def index(self):
|
||||||
|
"""Returns the audio layout as integer"""
|
||||||
|
return int(self.value['index'])
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def fromLabel(label : str):
|
||||||
|
try:
|
||||||
|
|
||||||
|
return [a for a in AudioLayout if a.value['label'] == str(label)][0]
|
||||||
|
except:
|
||||||
|
return AudioLayout.LAYOUT_UNDEFINED
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def fromIndex(index : int):
|
||||||
|
try:
|
||||||
|
return [a for a in AudioLayout if a.value['index'] == int(index)][0]
|
||||||
|
except:
|
||||||
|
return AudioLayout.LAYOUT_UNDEFINED
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def identify(streamObj):
|
||||||
|
|
||||||
|
FFPROBE_LAYOUT_KEY = 'channel_layout'
|
||||||
|
FFPROBE_CHANNELS_KEY = 'channels'
|
||||||
|
FFPROBE_CODEC_TYPE_KEY = 'codec_type'
|
||||||
|
|
||||||
|
if (type(streamObj) is not dict
|
||||||
|
or FFPROBE_CODEC_TYPE_KEY not in streamObj.keys()
|
||||||
|
or streamObj[FFPROBE_CODEC_TYPE_KEY] != TrackType.AUDIO.label()):
|
||||||
|
raise Exception('Not an ffprobe audio stream object')
|
||||||
|
|
||||||
|
if FFPROBE_LAYOUT_KEY in streamObj.keys():
|
||||||
|
matchingLayouts = [l for l in AudioLayout if l.label() == streamObj[FFPROBE_LAYOUT_KEY]]
|
||||||
|
if matchingLayouts:
|
||||||
|
return matchingLayouts[0]
|
||||||
|
|
||||||
|
if (FFPROBE_CHANNELS_KEY in streamObj.keys()
|
||||||
|
and int(streamObj[FFPROBE_CHANNELS_KEY]) == 6):
|
||||||
|
|
||||||
|
return AudioLayout.LAYOUT_6CH
|
||||||
|
|
||||||
|
return AudioLayout.LAYOUT_UNDEFINED
|
||||||
142
src/ffx/configuration_controller.py
Normal file
142
src/ffx/configuration_controller.py
Normal file
@@ -0,0 +1,142 @@
|
|||||||
|
import os, json
|
||||||
|
|
||||||
|
class ConfigurationController():
|
||||||
|
|
||||||
|
CONFIG_FILENAME = 'ffx.json'
|
||||||
|
DATABASE_FILENAME = 'ffx.db'
|
||||||
|
LOG_FILENAME = 'ffx.log'
|
||||||
|
|
||||||
|
DATABASE_PATH_CONFIG_KEY = 'databasePath'
|
||||||
|
LOG_DIRECTORY_CONFIG_KEY = 'logDirectory'
|
||||||
|
OUTPUT_FILENAME_TEMPLATE_KEY = 'outputFilenameTemplate'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
|
||||||
|
self.__homeDir = os.path.expanduser("~")
|
||||||
|
self.__localVarDir = os.path.join(self.__homeDir, '.local', 'var')
|
||||||
|
self.__localEtcDir = os.path.join(self.__homeDir, '.local', 'etc')
|
||||||
|
|
||||||
|
self.__configurationData = {}
|
||||||
|
|
||||||
|
# .local/etc/ffx.json
|
||||||
|
self.__configFilePath = os.path.join(self.__localEtcDir, ConfigurationController.CONFIG_FILENAME)
|
||||||
|
if os.path.isfile(self.__configFilePath):
|
||||||
|
with open(self.__configFilePath, 'r') as configurationFile:
|
||||||
|
self.__configurationData = json.load(configurationFile)
|
||||||
|
|
||||||
|
if ConfigurationController.DATABASE_PATH_CONFIG_KEY in self.__configurationData.keys():
|
||||||
|
self.__databaseFilePath = self.__configurationData[ConfigurationController.DATABASE_PATH_CONFIG_KEY]
|
||||||
|
os.makedirs(os.path.dirname(self.__databaseFilePath), exist_ok=True)
|
||||||
|
else:
|
||||||
|
ffxVarDir = os.path.join(self.__localVarDir, 'ffx')
|
||||||
|
os.makedirs(ffxVarDir, exist_ok=True)
|
||||||
|
self.__databaseFilePath = os.path.join(ffxVarDir, ConfigurationController.DATABASE_FILENAME)
|
||||||
|
|
||||||
|
if ConfigurationController.LOG_DIRECTORY_CONFIG_KEY in self.__configurationData.keys():
|
||||||
|
self.__logDir = self.__configurationData[ConfigurationController.LOG_DIRECTORY_CONFIG_KEY]
|
||||||
|
else:
|
||||||
|
self.__logDir = os.path.join(self.__localVarDir, 'log')
|
||||||
|
os.makedirs(self.__logDir, exist_ok=True)
|
||||||
|
|
||||||
|
|
||||||
|
def getHomeDirectory(self):
|
||||||
|
return self.__homeDir
|
||||||
|
|
||||||
|
def getLogFilePath(self):
|
||||||
|
return os.path.join(self.__logDir, ConfigurationController.LOG_FILENAME)
|
||||||
|
|
||||||
|
def getDatabaseFilePath(self):
|
||||||
|
return self.__databaseFilePath
|
||||||
|
|
||||||
|
|
||||||
|
def getData(self):
|
||||||
|
return self.__configurationData
|
||||||
|
|
||||||
|
|
||||||
|
#
|
||||||
|
#
|
||||||
|
#
|
||||||
|
# def addPattern(self, patternDescriptor):
|
||||||
|
#
|
||||||
|
# try:
|
||||||
|
#
|
||||||
|
# s = self.Session()
|
||||||
|
# q = s.query(Pattern).filter(Pattern.show_id == int(patternDescriptor['show_id']),
|
||||||
|
# Pattern.pattern == str(patternDescriptor['pattern']))
|
||||||
|
#
|
||||||
|
# if not q.count():
|
||||||
|
# pattern = Pattern(show_id = int(patternDescriptor['show_id']),
|
||||||
|
# pattern = str(patternDescriptor['pattern']))
|
||||||
|
# s.add(pattern)
|
||||||
|
# s.commit()
|
||||||
|
# return pattern.getId()
|
||||||
|
# else:
|
||||||
|
# return 0
|
||||||
|
#
|
||||||
|
# except Exception as ex:
|
||||||
|
# raise click.ClickException(f"PatternController.addPattern(): {repr(ex)}")
|
||||||
|
# finally:
|
||||||
|
# s.close()
|
||||||
|
#
|
||||||
|
#
|
||||||
|
# def updatePattern(self, patternId, patternDescriptor):
|
||||||
|
#
|
||||||
|
# try:
|
||||||
|
# s = self.Session()
|
||||||
|
# q = s.query(Pattern).filter(Pattern.id == int(patternId))
|
||||||
|
#
|
||||||
|
# if q.count():
|
||||||
|
#
|
||||||
|
# pattern = q.first()
|
||||||
|
#
|
||||||
|
# pattern.show_id = int(patternDescriptor['show_id'])
|
||||||
|
# pattern.pattern = str(patternDescriptor['pattern'])
|
||||||
|
#
|
||||||
|
# s.commit()
|
||||||
|
# return True
|
||||||
|
#
|
||||||
|
# else:
|
||||||
|
# return False
|
||||||
|
#
|
||||||
|
# except Exception as ex:
|
||||||
|
# raise click.ClickException(f"PatternController.updatePattern(): {repr(ex)}")
|
||||||
|
# finally:
|
||||||
|
# s.close()
|
||||||
|
#
|
||||||
|
#
|
||||||
|
#
|
||||||
|
# def findPattern(self, patternDescriptor):
|
||||||
|
#
|
||||||
|
# try:
|
||||||
|
# s = self.Session()
|
||||||
|
# q = s.query(Pattern).filter(Pattern.show_id == int(patternDescriptor['show_id']), Pattern.pattern == str(patternDescriptor['pattern']))
|
||||||
|
#
|
||||||
|
# if q.count():
|
||||||
|
# pattern = q.first()
|
||||||
|
# return int(pattern.id)
|
||||||
|
# else:
|
||||||
|
# return None
|
||||||
|
#
|
||||||
|
# except Exception as ex:
|
||||||
|
# raise click.ClickException(f"PatternController.findPattern(): {repr(ex)}")
|
||||||
|
# finally:
|
||||||
|
# s.close()
|
||||||
|
#
|
||||||
|
#
|
||||||
|
# def getPattern(self, patternId : int):
|
||||||
|
#
|
||||||
|
# if type(patternId) is not int:
|
||||||
|
# raise ValueError(f"PatternController.getPattern(): Argument patternId is required to be of type int")
|
||||||
|
#
|
||||||
|
# try:
|
||||||
|
# s = self.Session()
|
||||||
|
# q = s.query(Pattern).filter(Pattern.id == int(patternId))
|
||||||
|
#
|
||||||
|
# return q.first() if q.count() else None
|
||||||
|
#
|
||||||
|
# except Exception as ex:
|
||||||
|
# raise click.ClickException(f"PatternController.getPattern(): {repr(ex)}")
|
||||||
|
# finally:
|
||||||
|
# s.close()
|
||||||
|
#
|
||||||
15
src/ffx/constants.py
Normal file
15
src/ffx/constants.py
Normal file
@@ -0,0 +1,15 @@
|
|||||||
|
VERSION='0.2.3'
|
||||||
|
DATABASE_VERSION = 2
|
||||||
|
|
||||||
|
DEFAULT_QUALITY = 32
|
||||||
|
DEFAULT_AV1_PRESET = 5
|
||||||
|
|
||||||
|
DEFAULT_STEREO_BANDWIDTH = "112"
|
||||||
|
DEFAULT_AC3_BANDWIDTH = "256"
|
||||||
|
DEFAULT_DTS_BANDWIDTH = "320"
|
||||||
|
DEFAULT_7_1_BANDWIDTH = "384"
|
||||||
|
|
||||||
|
DEFAULT_CROP_START = 60
|
||||||
|
DEFAULT_CROP_LENGTH = 180
|
||||||
|
|
||||||
|
DEFAULT_OUTPUT_FILENAME_TEMPLATE = '{{ ffx_show_name }} - {{ ffx_index }}{{ ffx_index_separator }}{{ ffx_episode_name }}{{ ffx_indicator_separator }}{{ ffx_indicator }}'
|
||||||
102
src/ffx/database.py
Normal file
102
src/ffx/database.py
Normal file
@@ -0,0 +1,102 @@
|
|||||||
|
import os, click
|
||||||
|
|
||||||
|
from sqlalchemy import create_engine
|
||||||
|
from sqlalchemy.orm import sessionmaker
|
||||||
|
|
||||||
|
from ffx.model.show import Base
|
||||||
|
|
||||||
|
from ffx.model.property import Property
|
||||||
|
|
||||||
|
from ffx.constants import DATABASE_VERSION
|
||||||
|
|
||||||
|
|
||||||
|
DATABASE_VERSION_KEY = 'database_version'
|
||||||
|
|
||||||
|
class DatabaseVersionException(Exception):
|
||||||
|
def __init__(self, errorMessage):
|
||||||
|
super().__init__(errorMessage)
|
||||||
|
|
||||||
|
def databaseContext(databasePath: str = ''):
|
||||||
|
|
||||||
|
databaseContext = {}
|
||||||
|
|
||||||
|
if databasePath is None:
|
||||||
|
# sqlite:///:memory:
|
||||||
|
databasePath = ':memory:'
|
||||||
|
elif not databasePath:
|
||||||
|
homeDir = os.path.expanduser("~")
|
||||||
|
ffxVarDir = os.path.join(homeDir, '.local', 'var', 'ffx')
|
||||||
|
if not os.path.exists(ffxVarDir):
|
||||||
|
os.makedirs(ffxVarDir)
|
||||||
|
databasePath = os.path.join(ffxVarDir, 'ffx.db')
|
||||||
|
|
||||||
|
databaseContext['url'] = f"sqlite:///{databasePath}"
|
||||||
|
databaseContext['engine'] = create_engine(databaseContext['url'])
|
||||||
|
databaseContext['session'] = sessionmaker(bind=databaseContext['engine'])
|
||||||
|
|
||||||
|
Base.metadata.create_all(databaseContext['engine'])
|
||||||
|
|
||||||
|
# isSyncronuous = False
|
||||||
|
# while not isSyncronuous:
|
||||||
|
# while True:
|
||||||
|
# try:
|
||||||
|
# with databaseContext['database_engine'].connect() as connection:
|
||||||
|
# connection.execute(sqlalchemy.text('PRAGMA foreign_keys=ON;'))
|
||||||
|
# #isSyncronuous = True
|
||||||
|
# break
|
||||||
|
# except sqlite3.OperationalError:
|
||||||
|
# time.sleep(0.1)
|
||||||
|
|
||||||
|
ensureDatabaseVersion(databaseContext)
|
||||||
|
|
||||||
|
return databaseContext
|
||||||
|
|
||||||
|
def ensureDatabaseVersion(databaseContext):
|
||||||
|
|
||||||
|
currentDatabaseVersion = getDatabaseVersion(databaseContext)
|
||||||
|
if currentDatabaseVersion:
|
||||||
|
if currentDatabaseVersion != DATABASE_VERSION:
|
||||||
|
raise DatabaseVersionException(f"Current database version ({currentDatabaseVersion}) does not match required ({DATABASE_VERSION})")
|
||||||
|
else:
|
||||||
|
setDatabaseVersion(databaseContext, DATABASE_VERSION)
|
||||||
|
|
||||||
|
|
||||||
|
def getDatabaseVersion(databaseContext):
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
Session = databaseContext['session']
|
||||||
|
s = Session()
|
||||||
|
q = s.query(Property).filter(Property.key == DATABASE_VERSION_KEY)
|
||||||
|
|
||||||
|
return int(q.first().value) if q.count() else 0
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"getDatabaseVersion(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def setDatabaseVersion(databaseContext, databaseVersion: int):
|
||||||
|
|
||||||
|
try:
|
||||||
|
Session = databaseContext['session']
|
||||||
|
s = Session()
|
||||||
|
|
||||||
|
q = s.query(Property).filter(Property.key == DATABASE_VERSION_KEY)
|
||||||
|
|
||||||
|
dbVersion = int(databaseVersion)
|
||||||
|
|
||||||
|
versionProperty = q.first()
|
||||||
|
if versionProperty:
|
||||||
|
versionProperty.value = str(dbVersion)
|
||||||
|
else:
|
||||||
|
versionProperty = Property(key = DATABASE_VERSION_KEY,
|
||||||
|
value = str(dbVersion))
|
||||||
|
s.add(versionProperty)
|
||||||
|
s.commit()
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"setDatabaseVersion(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
740
src/ffx/ffx.py
Executable file
740
src/ffx/ffx.py
Executable file
@@ -0,0 +1,740 @@
|
|||||||
|
#! /usr/bin/python3
|
||||||
|
|
||||||
|
import os, click, time, logging
|
||||||
|
|
||||||
|
from ffx.configuration_controller import ConfigurationController
|
||||||
|
|
||||||
|
from ffx.file_properties import FileProperties
|
||||||
|
|
||||||
|
from ffx.ffx_app import FfxApp
|
||||||
|
from ffx.ffx_controller import FfxController
|
||||||
|
from ffx.tmdb_controller import TmdbController
|
||||||
|
|
||||||
|
from ffx.database import databaseContext
|
||||||
|
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.show_descriptor import ShowDescriptor
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
from ffx.video_encoder import VideoEncoder
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from ffx.track_codec import TrackCodec
|
||||||
|
|
||||||
|
from ffx.process import executeProcess
|
||||||
|
from ffx.helper import filterFilename, substituteTmdbFilename
|
||||||
|
from ffx.helper import getEpisodeFileBasename
|
||||||
|
|
||||||
|
from ffx.constants import DEFAULT_STEREO_BANDWIDTH, DEFAULT_AC3_BANDWIDTH, DEFAULT_DTS_BANDWIDTH, DEFAULT_7_1_BANDWIDTH
|
||||||
|
|
||||||
|
from ffx.filter.quality_filter import QualityFilter
|
||||||
|
from ffx.filter.preset_filter import PresetFilter
|
||||||
|
|
||||||
|
from ffx.filter.nlmeans_filter import NlmeansFilter
|
||||||
|
|
||||||
|
from ffx.constants import VERSION
|
||||||
|
|
||||||
|
from ffx.shifted_season_controller import ShiftedSeasonController
|
||||||
|
|
||||||
|
|
||||||
|
@click.group()
|
||||||
|
@click.pass_context
|
||||||
|
@click.option('--database-file', type=str, default='', help='Path to database file')
|
||||||
|
@click.option('-v', '--verbose', type=int, default=0, help='Set verbosity of output')
|
||||||
|
@click.option("--dry-run", is_flag=True, default=False)
|
||||||
|
def ffx(ctx, database_file, verbose, dry_run):
|
||||||
|
"""FFX"""
|
||||||
|
|
||||||
|
ctx.obj = {}
|
||||||
|
|
||||||
|
ctx.obj['config'] = ConfigurationController()
|
||||||
|
|
||||||
|
ctx.obj['database'] = databaseContext(databasePath=database_file
|
||||||
|
if database_file else ctx.obj['config'].getDatabaseFilePath())
|
||||||
|
|
||||||
|
ctx.obj['dry_run'] = dry_run
|
||||||
|
ctx.obj['verbosity'] = verbose
|
||||||
|
|
||||||
|
# Critical 50
|
||||||
|
# Error 40
|
||||||
|
# Warning 30
|
||||||
|
# Info 20
|
||||||
|
# Debug 10
|
||||||
|
fileLogVerbosity = max(40 - verbose * 10, 10)
|
||||||
|
consoleLogVerbosity = max(20 - verbose * 10, 10)
|
||||||
|
|
||||||
|
ctx.obj['logger'] = logging.getLogger('FFX')
|
||||||
|
ctx.obj['logger'].setLevel(logging.DEBUG)
|
||||||
|
|
||||||
|
ffxFileHandler = logging.FileHandler(ctx.obj['config'].getLogFilePath())
|
||||||
|
ffxFileHandler.setLevel(fileLogVerbosity)
|
||||||
|
ffxConsoleHandler = logging.StreamHandler()
|
||||||
|
ffxConsoleHandler.setLevel(consoleLogVerbosity)
|
||||||
|
|
||||||
|
fileFormatter = logging.Formatter(
|
||||||
|
'%(asctime)s - %(name)s - %(levelname)s - %(message)s')
|
||||||
|
ffxFileHandler.setFormatter(fileFormatter)
|
||||||
|
consoleFormatter = logging.Formatter(
|
||||||
|
'%(message)s')
|
||||||
|
ffxConsoleHandler.setFormatter(consoleFormatter)
|
||||||
|
|
||||||
|
ctx.obj['logger'].addHandler(ffxConsoleHandler)
|
||||||
|
ctx.obj['logger'].addHandler(ffxFileHandler)
|
||||||
|
|
||||||
|
|
||||||
|
# Define a subcommand
|
||||||
|
@ffx.command()
|
||||||
|
def version():
|
||||||
|
click.echo(VERSION)
|
||||||
|
|
||||||
|
|
||||||
|
# Another subcommand
|
||||||
|
@ffx.command()
|
||||||
|
def help():
|
||||||
|
click.echo(f"ffx {VERSION}\n")
|
||||||
|
click.echo(f"Usage: ffx [input file] [output file] [vp9|av1] [q=[nn[,nn,...]]] [p=nn] [a=nnn[k]] [ac3=nnn[k]] [dts=nnn[k]] [crop]")
|
||||||
|
|
||||||
|
|
||||||
|
@ffx.command()
|
||||||
|
@click.pass_context
|
||||||
|
@click.argument('filename', nargs=1)
|
||||||
|
def inspect(ctx, filename):
|
||||||
|
|
||||||
|
ctx.obj['command'] = 'inspect'
|
||||||
|
ctx.obj['arguments'] = {}
|
||||||
|
ctx.obj['arguments']['filename'] = filename
|
||||||
|
|
||||||
|
app = FfxApp(ctx.obj)
|
||||||
|
app.run()
|
||||||
|
|
||||||
|
|
||||||
|
def getUnmuxSequence(trackDescriptor: TrackDescriptor, sourcePath, targetPrefix, targetDirectory = ''):
|
||||||
|
|
||||||
|
# executable and input file
|
||||||
|
commandTokens = FfxController.COMMAND_TOKENS + ['-i', sourcePath]
|
||||||
|
|
||||||
|
trackType = trackDescriptor.getType()
|
||||||
|
|
||||||
|
targetPathBase = os.path.join(targetDirectory, targetPrefix) if targetDirectory else targetPrefix
|
||||||
|
|
||||||
|
# mapping
|
||||||
|
commandTokens += ['-map',
|
||||||
|
f"0:{trackType.indicator()}:{trackDescriptor.getSubIndex()}",
|
||||||
|
'-c',
|
||||||
|
'copy']
|
||||||
|
|
||||||
|
trackCodec = trackDescriptor.getCodec()
|
||||||
|
|
||||||
|
# output format
|
||||||
|
codecFormat = trackCodec.format()
|
||||||
|
if codecFormat is not None:
|
||||||
|
commandTokens += ['-f', codecFormat]
|
||||||
|
|
||||||
|
# output filename
|
||||||
|
commandTokens += [f"{targetPathBase}.{trackCodec.extension()}"]
|
||||||
|
|
||||||
|
return commandTokens
|
||||||
|
|
||||||
|
|
||||||
|
@ffx.command()
|
||||||
|
@click.pass_context
|
||||||
|
|
||||||
|
@click.argument('paths', nargs=-1)
|
||||||
|
@click.option('-l', '--label', type=str, default='', help='Label to be used as filename prefix')
|
||||||
|
@click.option("-o", "--output-directory", type=str, default='')
|
||||||
|
@click.option("-s", "--subtitles-only", is_flag=True, default=False)
|
||||||
|
@click.option('--nice', type=int, default=99, help='Niceness of started processes')
|
||||||
|
@click.option('--cpu', type=int, default=0, help='Limit CPU for started processes to percent')
|
||||||
|
def unmux(ctx,
|
||||||
|
paths,
|
||||||
|
label,
|
||||||
|
output_directory,
|
||||||
|
subtitles_only,
|
||||||
|
nice,
|
||||||
|
cpu):
|
||||||
|
|
||||||
|
existingSourcePaths = [p for p in paths if os.path.isfile(p)]
|
||||||
|
ctx.obj['logger'].debug(f"\nUnmuxing {len(existingSourcePaths)} files")
|
||||||
|
|
||||||
|
ctx.obj['resource_limits'] = {}
|
||||||
|
ctx.obj['resource_limits']['niceness'] = nice
|
||||||
|
ctx.obj['resource_limits']['cpu_percent'] = cpu
|
||||||
|
|
||||||
|
for sourcePath in existingSourcePaths:
|
||||||
|
|
||||||
|
fp = FileProperties(ctx.obj, sourcePath)
|
||||||
|
|
||||||
|
|
||||||
|
try:
|
||||||
|
sourceMediaDescriptor = fp.getMediaDescriptor()
|
||||||
|
|
||||||
|
season = fp.getSeason()
|
||||||
|
episode = fp.getEpisode()
|
||||||
|
|
||||||
|
#TODO: Recognition für alle Formate anpassen
|
||||||
|
targetLabel = label if label else fp.getFileBasename()
|
||||||
|
targetIndicator = f"_S{season}E{episode}" if label and season != -1 and episode != -1 else ''
|
||||||
|
|
||||||
|
if label and not targetIndicator:
|
||||||
|
ctx.obj['logger'].warning(f"Skipping file {fp.getFilename()}: Label set but no indicator recognized")
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
ctx.obj['logger'].info(f"\nUnmuxing file {fp.getFilename()}\n")
|
||||||
|
|
||||||
|
for trackDescriptor in sourceMediaDescriptor.getAllTrackDescriptors():
|
||||||
|
|
||||||
|
if trackDescriptor.getType() == TrackType.SUBTITLE or not subtitles_only:
|
||||||
|
|
||||||
|
# SEASON_EPISODE_STREAM_LANGUAGE_MATCH = '[sS]([0-9]+)[eE]([0-9]+)_([0-9]+)_([a-z]{3})(?:_([A-Z]{3}))*'
|
||||||
|
targetPrefix = f"{targetLabel}{targetIndicator}_{trackDescriptor.getIndex()}_{trackDescriptor.getLanguage().threeLetter()}"
|
||||||
|
|
||||||
|
td: TrackDisposition
|
||||||
|
for td in sorted(trackDescriptor.getDispositionSet(), key=lambda d: d.index()):
|
||||||
|
targetPrefix += f"_{td.indicator()}"
|
||||||
|
|
||||||
|
unmuxSequence = getUnmuxSequence(trackDescriptor, sourcePath, targetPrefix, targetDirectory = output_directory)
|
||||||
|
|
||||||
|
if unmuxSequence:
|
||||||
|
if not ctx.obj['dry_run']:
|
||||||
|
|
||||||
|
#TODO #425: Codec Enum
|
||||||
|
ctx.obj['logger'].info(f"Unmuxing stream {trackDescriptor.getIndex()} into file {targetPrefix}.{trackDescriptor.getCodec().extension()}")
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"Executing unmuxing sequence")
|
||||||
|
|
||||||
|
out, err, rc = executeProcess(unmuxSequence, context = ctx.obj)
|
||||||
|
if rc:
|
||||||
|
ctx.obj['logger'].error(f"Unmuxing of stream {trackDescriptor.getIndex()} failed with error ({rc}) {err}")
|
||||||
|
else:
|
||||||
|
ctx.obj['logger'].warning(f"Skipping stream with unknown codec")
|
||||||
|
except Exception as ex:
|
||||||
|
ctx.obj['logger'].warning(f"Skipping File {sourcePath} ({ex})")
|
||||||
|
|
||||||
|
|
||||||
|
@ffx.command()
|
||||||
|
@click.pass_context
|
||||||
|
|
||||||
|
def shows(ctx):
|
||||||
|
|
||||||
|
ctx.obj['command'] = 'shows'
|
||||||
|
|
||||||
|
app = FfxApp(ctx.obj)
|
||||||
|
app.run()
|
||||||
|
|
||||||
|
|
||||||
|
def checkUniqueDispositions(context, mediaDescriptor: MediaDescriptor):
|
||||||
|
|
||||||
|
# Check for multiple default or forced dispositions if not set by user input or database requirements
|
||||||
|
#
|
||||||
|
# Query user for the correct sub indices, then configure flags in track descriptors associated with media descriptor accordingly.
|
||||||
|
# The correct tokens should then be created by
|
||||||
|
if len([v for v in mediaDescriptor.getVideoTracks() if v.getDispositionFlag(TrackDisposition.DEFAULT)]) > 1:
|
||||||
|
if context['no_prompt']:
|
||||||
|
raise click.ClickException('More than one default video stream detected and no prompt set')
|
||||||
|
defaultVideoTrackSubIndex = click.prompt("More than one default video stream detected! Please select stream", type=int)
|
||||||
|
mediaDescriptor.setDefaultSubTrack(TrackType.VIDEO, defaultVideoTrackSubIndex)
|
||||||
|
|
||||||
|
if len([v for v in mediaDescriptor.getVideoTracks() if v.getDispositionFlag(TrackDisposition.FORCED)]) > 1:
|
||||||
|
if context['no_prompt']:
|
||||||
|
raise click.ClickException('More than one forced video stream detected and no prompt set')
|
||||||
|
forcedVideoTrackSubIndex = click.prompt("More than one forced video stream detected! Please select stream", type=int)
|
||||||
|
mediaDescriptor.setForcedSubTrack(TrackType.VIDEO, forcedVideoTrackSubIndex)
|
||||||
|
|
||||||
|
if len([a for a in mediaDescriptor.getAudioTracks() if a.getDispositionFlag(TrackDisposition.DEFAULT)]) > 1:
|
||||||
|
if context['no_prompt']:
|
||||||
|
raise click.ClickException('More than one default audio stream detected and no prompt set')
|
||||||
|
defaultAudioTrackSubIndex = click.prompt("More than one default audio stream detected! Please select stream", type=int)
|
||||||
|
mediaDescriptor.setDefaultSubTrack(TrackType.AUDIO, defaultAudioTrackSubIndex)
|
||||||
|
|
||||||
|
if len([a for a in mediaDescriptor.getAudioTracks() if a.getDispositionFlag(TrackDisposition.FORCED)]) > 1:
|
||||||
|
if context['no_prompt']:
|
||||||
|
raise click.ClickException('More than one forced audio stream detected and no prompt set')
|
||||||
|
forcedAudioTrackSubIndex = click.prompt("More than one forced audio stream detected! Please select stream", type=int)
|
||||||
|
mediaDescriptor.setForcedSubTrack(TrackType.AUDIO, forcedAudioTrackSubIndex)
|
||||||
|
|
||||||
|
if len([s for s in mediaDescriptor.getSubtitleTracks() if s.getDispositionFlag(TrackDisposition.DEFAULT)]) > 1:
|
||||||
|
if context['no_prompt']:
|
||||||
|
raise click.ClickException('More than one default subtitle stream detected and no prompt set')
|
||||||
|
defaultSubtitleTrackSubIndex = click.prompt("More than one default subtitle stream detected! Please select stream", type=int)
|
||||||
|
mediaDescriptor.setDefaultSubTrack(TrackType.SUBTITLE, defaultSubtitleTrackSubIndex)
|
||||||
|
|
||||||
|
if len([s for s in mediaDescriptor.getSubtitleTracks() if s.getDispositionFlag(TrackDisposition.FORCED)]) > 1:
|
||||||
|
if context['no_prompt']:
|
||||||
|
raise click.ClickException('More than one forced subtitle stream detected and no prompt set')
|
||||||
|
forcedSubtitleTrackSubIndex = click.prompt("More than one forced subtitle stream detected! Please select stream", type=int)
|
||||||
|
mediaDescriptor.setForcedSubTrack(TrackType.SUBTITLE, forcedSubtitleTrackSubIndex)
|
||||||
|
|
||||||
|
|
||||||
|
@ffx.command()
|
||||||
|
@click.pass_context
|
||||||
|
|
||||||
|
@click.argument('paths', nargs=-1)
|
||||||
|
|
||||||
|
@click.option('-l', '--label', type=str, default='', help='Label to be used as filename prefix')
|
||||||
|
|
||||||
|
@click.option('-v', '--video-encoder', type=str, default=FfxController.DEFAULT_VIDEO_ENCODER, help=f"Target video encoder (vp9 or av1)", show_default=True)
|
||||||
|
|
||||||
|
@click.option('-q', '--quality', type=str, default="", help=f"Quality settings to be used with VP9 encoder")
|
||||||
|
@click.option('-p', '--preset', type=str, default="", help=f"Quality preset to be used with AV1 encoder")
|
||||||
|
|
||||||
|
@click.option('-a', '--stereo-bitrate', type=int, default=DEFAULT_STEREO_BANDWIDTH, help=f"Bitrate in kbit/s to be used to encode stereo audio streams", show_default=True)
|
||||||
|
@click.option('--ac3', type=int, default=DEFAULT_AC3_BANDWIDTH, help=f"Bitrate in kbit/s to be used to encode 5.1 audio streams", show_default=True)
|
||||||
|
@click.option('--dts', type=int, default=DEFAULT_DTS_BANDWIDTH, help=f"Bitrate in kbit/s to be used to encode 6.1 audio streams", show_default=True)
|
||||||
|
|
||||||
|
@click.option('--subtitle-directory', type=str, default='', help='Load subtitles from here')
|
||||||
|
@click.option('--subtitle-prefix', type=str, default='', help='Subtitle filename prefix')
|
||||||
|
|
||||||
|
@click.option('--language', type=str, multiple=True, help='Set stream language. Use format <stream index>:<3 letter iso code>')
|
||||||
|
@click.option('--title', type=str, multiple=True, help='Set stream title. Use format <stream index>:<title>')
|
||||||
|
|
||||||
|
@click.option('--default-video', type=int, default=-1, help='Index of default video stream')
|
||||||
|
@click.option('--forced-video', type=int, default=-1, help='Index of forced video stream')
|
||||||
|
@click.option('--default-audio', type=int, default=-1, help='Index of default audio stream')
|
||||||
|
@click.option('--forced-audio', type=int, default=-1, help='Index of forced audio stream')
|
||||||
|
@click.option('--default-subtitle', type=int, default=-1, help='Index of default subtitle stream')
|
||||||
|
@click.option('--forced-subtitle', type=int, default=-1, help='Index of forced subtitle stream')
|
||||||
|
|
||||||
|
@click.option('--rearrange-streams', type=str, default="", help='Rearrange output streams order. Use format comma separated integers')
|
||||||
|
|
||||||
|
@click.option("--crop", is_flag=False, flag_value="default", default="none")
|
||||||
|
|
||||||
|
@click.option("--output-directory", type=str, default='')
|
||||||
|
|
||||||
|
@click.option("--denoise", is_flag=False, flag_value="default", default="none")
|
||||||
|
@click.option("--denoise-use-hw", is_flag=True, default=False)
|
||||||
|
@click.option('--denoise-strength', type=str, default='', help='Denoising strength, more blurring vs more details.')
|
||||||
|
@click.option('--denoise-patch-size', type=str, default='', help='Subimage size to apply filtering on luminosity plane. Reduces broader noise patterns but costly.')
|
||||||
|
@click.option('--denoise-chroma-patch-size', type=str, default='', help='Subimage size to apply filtering on chroma planes.')
|
||||||
|
@click.option('--denoise-research-window', type=str, default='', help='Range to search for comparable patches on luminosity plane. Better filtering but costly.')
|
||||||
|
@click.option('--denoise-chroma-research-window', type=str, default='', help='Range to search for comparable patches on chroma planes.')
|
||||||
|
|
||||||
|
@click.option('--show', type=int, default=-1, help='Set TMDB show identifier')
|
||||||
|
@click.option('--season', type=int, default=-1, help='Set season of show')
|
||||||
|
@click.option('--episode', type=int, default=-1, help='Set episode of show')
|
||||||
|
|
||||||
|
@click.option("--no-tmdb", is_flag=True, default=False)
|
||||||
|
@click.option("--no-pattern", is_flag=True, default=False)
|
||||||
|
|
||||||
|
@click.option("--dont-pass-dispositions", is_flag=True, default=False)
|
||||||
|
|
||||||
|
@click.option("--no-prompt", is_flag=True, default=False)
|
||||||
|
@click.option("--no-signature", is_flag=True, default=False)
|
||||||
|
@click.option("--keep-mkvmerge-metadata", is_flag=True, default=False)
|
||||||
|
|
||||||
|
@click.option('--nice', type=int, default=99, help='Niceness of started processes')
|
||||||
|
@click.option('--cpu', type=int, default=0, help='Limit CPU for started processes to percent')
|
||||||
|
|
||||||
|
def convert(ctx,
|
||||||
|
paths,
|
||||||
|
label,
|
||||||
|
video_encoder,
|
||||||
|
quality,
|
||||||
|
preset,
|
||||||
|
stereo_bitrate,
|
||||||
|
ac3,
|
||||||
|
dts,
|
||||||
|
|
||||||
|
subtitle_directory,
|
||||||
|
subtitle_prefix,
|
||||||
|
|
||||||
|
language,
|
||||||
|
title,
|
||||||
|
|
||||||
|
default_video,
|
||||||
|
forced_video,
|
||||||
|
default_audio,
|
||||||
|
forced_audio,
|
||||||
|
default_subtitle,
|
||||||
|
forced_subtitle,
|
||||||
|
|
||||||
|
rearrange_streams,
|
||||||
|
|
||||||
|
crop,
|
||||||
|
output_directory,
|
||||||
|
|
||||||
|
denoise,
|
||||||
|
denoise_use_hw,
|
||||||
|
denoise_strength,
|
||||||
|
denoise_patch_size,
|
||||||
|
denoise_chroma_patch_size,
|
||||||
|
denoise_research_window,
|
||||||
|
denoise_chroma_research_window,
|
||||||
|
|
||||||
|
show,
|
||||||
|
season,
|
||||||
|
episode,
|
||||||
|
|
||||||
|
no_tmdb,
|
||||||
|
no_pattern,
|
||||||
|
dont_pass_dispositions,
|
||||||
|
no_prompt,
|
||||||
|
no_signature,
|
||||||
|
keep_mkvmerge_metadata,
|
||||||
|
|
||||||
|
nice,
|
||||||
|
cpu):
|
||||||
|
"""Batch conversion of audiovideo files in format suitable for web playback, e.g. jellyfin
|
||||||
|
|
||||||
|
Files found under PATHS will be converted according to parameters.
|
||||||
|
Filename extensions will be changed appropriately.
|
||||||
|
Suffices will we appended to filename in case of multiple created files
|
||||||
|
or if the filename has not changed."""
|
||||||
|
|
||||||
|
startTime = time.perf_counter()
|
||||||
|
|
||||||
|
context = ctx.obj
|
||||||
|
|
||||||
|
context['video_encoder'] = VideoEncoder.fromLabel(video_encoder)
|
||||||
|
|
||||||
|
targetFormat = FfxController.DEFAULT_FILE_FORMAT
|
||||||
|
targetExtension = FfxController.DEFAULT_FILE_EXTENSION
|
||||||
|
|
||||||
|
context['use_tmdb'] = not no_tmdb
|
||||||
|
context['use_pattern'] = not no_pattern
|
||||||
|
context['no_prompt'] = no_prompt
|
||||||
|
context['no_signature'] = no_signature
|
||||||
|
context['keep_mkvmerge_metadata'] = keep_mkvmerge_metadata
|
||||||
|
|
||||||
|
|
||||||
|
context['resource_limits'] = {}
|
||||||
|
context['resource_limits']['niceness'] = nice
|
||||||
|
context['resource_limits']['cpu_percent'] = cpu
|
||||||
|
|
||||||
|
|
||||||
|
context['import_subtitles'] = (subtitle_directory and subtitle_prefix)
|
||||||
|
if context['import_subtitles']:
|
||||||
|
context['subtitle_directory'] = subtitle_directory
|
||||||
|
context['subtitle_prefix'] = subtitle_prefix
|
||||||
|
|
||||||
|
|
||||||
|
existingSourcePaths = [p for p in paths if os.path.isfile(p) and p.split('.')[-1] in FfxController.INPUT_FILE_EXTENSIONS]
|
||||||
|
|
||||||
|
|
||||||
|
# CLI Overrides
|
||||||
|
|
||||||
|
cliOverrides = {}
|
||||||
|
|
||||||
|
if language:
|
||||||
|
cliOverrides['languages'] = {}
|
||||||
|
for overLang in language:
|
||||||
|
olTokens = overLang.split(':')
|
||||||
|
if len(olTokens) == 2:
|
||||||
|
try:
|
||||||
|
cliOverrides['languages'][int(olTokens[0])] = olTokens[1]
|
||||||
|
except ValueError:
|
||||||
|
ctx.obj['logger'].warning(f"Ignoring non-integer language index {olTokens[0]}")
|
||||||
|
continue
|
||||||
|
|
||||||
|
if title:
|
||||||
|
cliOverrides['titles'] = {}
|
||||||
|
for overTitle in title:
|
||||||
|
otTokens = overTitle.split(':')
|
||||||
|
if len(otTokens) == 2:
|
||||||
|
try:
|
||||||
|
cliOverrides['titles'][int(otTokens[0])] = otTokens[1]
|
||||||
|
except ValueError:
|
||||||
|
ctx.obj['logger'].warning(f"Ignoring non-integer title index {otTokens[0]}")
|
||||||
|
continue
|
||||||
|
|
||||||
|
if default_video != -1:
|
||||||
|
cliOverrides['default_video'] = default_video
|
||||||
|
if forced_video != -1:
|
||||||
|
cliOverrides['forced_video'] = forced_video
|
||||||
|
if default_audio != -1:
|
||||||
|
cliOverrides['default_audio'] = default_audio
|
||||||
|
if forced_audio != -1:
|
||||||
|
cliOverrides['forced_audio'] = forced_audio
|
||||||
|
if default_subtitle != -1:
|
||||||
|
cliOverrides['default_subtitle'] = default_subtitle
|
||||||
|
if forced_subtitle != -1:
|
||||||
|
cliOverrides['forced_subtitle'] = forced_subtitle
|
||||||
|
|
||||||
|
if show != -1 or season != -1 or episode != -1:
|
||||||
|
if len(existingSourcePaths) > 1:
|
||||||
|
context['logger'].warning(f"Ignoring TMDB show, season, episode overrides, not supported for multiple source files")
|
||||||
|
else:
|
||||||
|
cliOverrides['tmdb'] = {}
|
||||||
|
if show != -1:
|
||||||
|
cliOverrides['tmdb']['show'] = show
|
||||||
|
if season != -1:
|
||||||
|
cliOverrides['tmdb']['season'] = season
|
||||||
|
if episode != -1:
|
||||||
|
cliOverrides['tmdb']['episode'] = episode
|
||||||
|
|
||||||
|
if cliOverrides:
|
||||||
|
context['overrides'] = cliOverrides
|
||||||
|
|
||||||
|
|
||||||
|
if rearrange_streams:
|
||||||
|
try:
|
||||||
|
cliOverrides['stream_order'] = [int(si) for si in rearrange_streams.split(",")]
|
||||||
|
except ValueError as ve:
|
||||||
|
errorMessage = "Non-integer in rearrange stream parameter"
|
||||||
|
ctx.obj['logger'].error(errorMessage)
|
||||||
|
raise click.Abort()
|
||||||
|
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"\nVideo encoder: {video_encoder}")
|
||||||
|
|
||||||
|
qualityTokens = quality.split(',')
|
||||||
|
q_list = [q for q in qualityTokens if q.isnumeric()]
|
||||||
|
ctx.obj['logger'].debug(f"Qualities: {q_list}")
|
||||||
|
|
||||||
|
presetTokens = preset.split(',')
|
||||||
|
p_list = [p for p in presetTokens if p.isnumeric()]
|
||||||
|
ctx.obj['logger'].debug(f"Presets: {p_list}")
|
||||||
|
|
||||||
|
|
||||||
|
context['bitrates'] = {}
|
||||||
|
context['bitrates']['stereo'] = str(stereo_bitrate) if str(stereo_bitrate).endswith('k') else f"{stereo_bitrate}k"
|
||||||
|
context['bitrates']['ac3'] = str(ac3) if str(ac3).endswith('k') else f"{ac3}k"
|
||||||
|
context['bitrates']['dts'] = str(dts) if str(dts).endswith('k') else f"{dts}k"
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"Stereo bitrate: {context['bitrates']['stereo']}")
|
||||||
|
ctx.obj['logger'].debug(f"AC3 bitrate: {context['bitrates']['ac3']}")
|
||||||
|
ctx.obj['logger'].debug(f"DTS bitrate: {context['bitrates']['dts']}")
|
||||||
|
|
||||||
|
|
||||||
|
# Process crop parameters
|
||||||
|
context['perform_crop'] = (crop != 'none')
|
||||||
|
if context['perform_crop']:
|
||||||
|
cTokens = crop.split(',')
|
||||||
|
if cTokens and len(cTokens) == 2:
|
||||||
|
context['crop_start'] = int(cTokens[0])
|
||||||
|
context['crop_length'] = int(cTokens[1])
|
||||||
|
ctx.obj['logger'].debug(f"Crop start={context['crop_start']} length={context['crop_length']}")
|
||||||
|
|
||||||
|
|
||||||
|
tc = TmdbController() if context['use_tmdb'] else None
|
||||||
|
|
||||||
|
qualityKwargs = {QualityFilter.QUALITY_KEY: quality}
|
||||||
|
qf = QualityFilter(**qualityKwargs)
|
||||||
|
|
||||||
|
if context['video_encoder'] == VideoEncoder.AV1 and preset:
|
||||||
|
presetKwargs = {PresetFilter.PRESET_KEY: preset}
|
||||||
|
PresetFilter(**presetKwargs)
|
||||||
|
|
||||||
|
denoiseKwargs = {}
|
||||||
|
if denoise_strength:
|
||||||
|
denoiseKwargs[NlmeansFilter.STRENGTH_KEY] = denoise_strength
|
||||||
|
if denoise_patch_size:
|
||||||
|
denoiseKwargs[NlmeansFilter.PATCH_SIZE_KEY] = denoise_patch_size
|
||||||
|
if denoise_chroma_patch_size:
|
||||||
|
denoiseKwargs[NlmeansFilter.CHROMA_PATCH_SIZE_KEY] = denoise_chroma_patch_size
|
||||||
|
if denoise_research_window:
|
||||||
|
denoiseKwargs[NlmeansFilter.RESEARCH_WINDOW_KEY] = denoise_research_window
|
||||||
|
if denoise_chroma_research_window:
|
||||||
|
denoiseKwargs[NlmeansFilter.CHROMA_RESEARCH_WINDOW_KEY] = denoise_chroma_research_window
|
||||||
|
if denoise != 'none' or denoiseKwargs:
|
||||||
|
NlmeansFilter(**denoiseKwargs)
|
||||||
|
|
||||||
|
chainYield = list(qf.getChainYield())
|
||||||
|
|
||||||
|
ctx.obj['logger'].info(f"\nRunning {len(existingSourcePaths) * len(chainYield)} jobs")
|
||||||
|
|
||||||
|
jobIndex = 0
|
||||||
|
|
||||||
|
for sourcePath in existingSourcePaths:
|
||||||
|
|
||||||
|
# Separate basedir, basename and extension for current source file
|
||||||
|
sourceDirectory = os.path.dirname(sourcePath)
|
||||||
|
sourceFilename = os.path.basename(sourcePath)
|
||||||
|
sourcePathTokens = sourceFilename.split('.')
|
||||||
|
|
||||||
|
sourceFileBasename = '.'.join(sourcePathTokens[:-1])
|
||||||
|
sourceFilenameExtension = sourcePathTokens[-1]
|
||||||
|
|
||||||
|
ctx.obj['logger'].info(f"\nProcessing file {sourcePath}")
|
||||||
|
|
||||||
|
targetSuffices = {}
|
||||||
|
|
||||||
|
|
||||||
|
mediaFileProperties = FileProperties(context, sourceFilename)
|
||||||
|
|
||||||
|
|
||||||
|
ssc = ShiftedSeasonController(context)
|
||||||
|
|
||||||
|
showId = mediaFileProperties.getShowId()
|
||||||
|
|
||||||
|
#HINT: -1 if not set
|
||||||
|
if 'tmdb' in cliOverrides.keys() and 'season' in cliOverrides['tmdb']:
|
||||||
|
showSeason = cliOverrides['tmdb']['season']
|
||||||
|
else:
|
||||||
|
showSeason = mediaFileProperties.getSeason()
|
||||||
|
|
||||||
|
if 'tmdb' in cliOverrides.keys() and 'episode' in cliOverrides['tmdb']:
|
||||||
|
showEpisode = cliOverrides['tmdb']['episode']
|
||||||
|
else:
|
||||||
|
showEpisode = mediaFileProperties.getEpisode()
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"Season={showSeason} Episode={showEpisode}")
|
||||||
|
|
||||||
|
|
||||||
|
sourceMediaDescriptor = mediaFileProperties.getMediaDescriptor()
|
||||||
|
|
||||||
|
#HINT: This is None if the filename did not match anything in database
|
||||||
|
currentPattern = mediaFileProperties.getPattern() if context['use_pattern'] else None
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"Pattern matching: {'No' if currentPattern is None else 'Yes'}")
|
||||||
|
|
||||||
|
# Setup FfxController accordingly depending on pattern matching is enabled and a pattern was matched
|
||||||
|
if currentPattern is None:
|
||||||
|
|
||||||
|
checkUniqueDispositions(context, sourceMediaDescriptor)
|
||||||
|
currentShowDescriptor = None
|
||||||
|
|
||||||
|
if context['import_subtitles']:
|
||||||
|
sourceMediaDescriptor.importSubtitles(context['subtitle_directory'],
|
||||||
|
context['subtitle_prefix'],
|
||||||
|
showSeason,
|
||||||
|
showEpisode)
|
||||||
|
|
||||||
|
if cliOverrides:
|
||||||
|
sourceMediaDescriptor.applyOverrides(cliOverrides)
|
||||||
|
|
||||||
|
fc = FfxController(context, sourceMediaDescriptor)
|
||||||
|
|
||||||
|
else:
|
||||||
|
targetMediaDescriptor = currentPattern.getMediaDescriptor(ctx.obj)
|
||||||
|
checkUniqueDispositions(context, targetMediaDescriptor)
|
||||||
|
currentShowDescriptor = currentPattern.getShowDescriptor(ctx.obj)
|
||||||
|
|
||||||
|
if context['import_subtitles']:
|
||||||
|
targetMediaDescriptor.importSubtitles(context['subtitle_directory'],
|
||||||
|
context['subtitle_prefix'],
|
||||||
|
showSeason,
|
||||||
|
showEpisode)
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"tmd subindices: {[t.getIndex() for t in targetMediaDescriptor.getAllTrackDescriptors()]} {[t.getSubIndex() for t in targetMediaDescriptor.getAllTrackDescriptors()]} {[t.getDispositionFlag(TrackDisposition.DEFAULT) for t in targetMediaDescriptor.getAllTrackDescriptors()]}")
|
||||||
|
|
||||||
|
if cliOverrides:
|
||||||
|
targetMediaDescriptor.applyOverrides(cliOverrides)
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"tmd subindices: {[t.getIndex() for t in targetMediaDescriptor.getAllTrackDescriptors()]} {[t.getSubIndex() for t in targetMediaDescriptor.getAllTrackDescriptors()]} {[t.getDispositionFlag(TrackDisposition.DEFAULT) for t in targetMediaDescriptor.getAllTrackDescriptors()]}")
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"Input mapping tokens (2nd pass): {targetMediaDescriptor.getInputMappingTokens()}")
|
||||||
|
|
||||||
|
fc = FfxController(context, targetMediaDescriptor, sourceMediaDescriptor)
|
||||||
|
|
||||||
|
|
||||||
|
indexSeasonDigits = currentShowDescriptor.getIndexSeasonDigits() if not currentPattern is None else ShowDescriptor.DEFAULT_INDEX_SEASON_DIGITS
|
||||||
|
indexEpisodeDigits = currentShowDescriptor.getIndexEpisodeDigits() if not currentPattern is None else ShowDescriptor.DEFAULT_INDEX_EPISODE_DIGITS
|
||||||
|
indicatorSeasonDigits = currentShowDescriptor.getIndicatorSeasonDigits() if not currentPattern is None else ShowDescriptor.DEFAULT_INDICATOR_SEASON_DIGITS
|
||||||
|
indicatorEpisodeDigits = currentShowDescriptor.getIndicatorEpisodeDigits() if not currentPattern is None else ShowDescriptor.DEFAULT_INDICATOR_EPISODE_DIGITS
|
||||||
|
|
||||||
|
|
||||||
|
# Shift season and episode if defined for this show
|
||||||
|
if ('tmdb' not in cliOverrides.keys() and showId != -1
|
||||||
|
and showSeason != -1 and showEpisode != -1):
|
||||||
|
shiftedShowSeason, shiftedShowEpisode = ssc.shiftSeason(showId,
|
||||||
|
season=showSeason,
|
||||||
|
episode=showEpisode)
|
||||||
|
else:
|
||||||
|
shiftedShowSeason = showSeason
|
||||||
|
shiftedShowEpisode = showEpisode
|
||||||
|
|
||||||
|
# Assemble target filename accordingly depending on TMDB lookup is enabled
|
||||||
|
#HINT: -1 if not set
|
||||||
|
showId = cliOverrides['tmdb']['show'] if 'tmdb' in cliOverrides.keys() and 'show' in cliOverrides['tmdb'] else (-1 if currentShowDescriptor is None else currentShowDescriptor.getId())
|
||||||
|
|
||||||
|
if context['use_tmdb'] and showId != -1 and shiftedShowSeason != -1 and shiftedShowEpisode != -1:
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"Querying TMDB for show_id={showId} season={shiftedShowSeason} episode{shiftedShowEpisode}")
|
||||||
|
|
||||||
|
if currentPattern is None:
|
||||||
|
sName, showYear = tc.getShowNameAndYear(showId)
|
||||||
|
showName = filterFilename(sName)
|
||||||
|
showFilenamePrefix = f"{showName} ({str(showYear)})"
|
||||||
|
else:
|
||||||
|
showFilenamePrefix = currentShowDescriptor.getFilenamePrefix()
|
||||||
|
|
||||||
|
tmdbEpisodeResult = tc.queryEpisode(showId, shiftedShowSeason, shiftedShowEpisode)
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"tmdbEpisodeResult={tmdbEpisodeResult}")
|
||||||
|
|
||||||
|
if tmdbEpisodeResult:
|
||||||
|
substitutedEpisodeName = filterFilename(substituteTmdbFilename(tmdbEpisodeResult['name']))
|
||||||
|
sourceFileBasename = getEpisodeFileBasename(showFilenamePrefix,
|
||||||
|
substitutedEpisodeName,
|
||||||
|
shiftedShowSeason,
|
||||||
|
shiftedShowEpisode,
|
||||||
|
indexSeasonDigits,
|
||||||
|
indexEpisodeDigits,
|
||||||
|
indicatorSeasonDigits,
|
||||||
|
indicatorEpisodeDigits,
|
||||||
|
context=ctx.obj)
|
||||||
|
|
||||||
|
if label:
|
||||||
|
if shiftedShowSeason > -1 and shiftedShowEpisode > -1:
|
||||||
|
targetSuffices['se'] = f"S{shiftedShowSeason:0{indicatorSeasonDigits}d}E{shiftedShowEpisode:0{indicatorEpisodeDigits}d}"
|
||||||
|
elif shiftedShowEpisode > -1:
|
||||||
|
targetSuffices['se'] = f"E{shiftedShowEpisode:0{indicatorEpisodeDigits}d}"
|
||||||
|
else:
|
||||||
|
if 'se' in targetSuffices.keys():
|
||||||
|
del targetSuffices['se']
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"fileBasename={sourceFileBasename}")
|
||||||
|
|
||||||
|
|
||||||
|
for chainIteration in chainYield:
|
||||||
|
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"\nchain iteration: {chainIteration}\n")
|
||||||
|
|
||||||
|
# if len(q_list) > 1:
|
||||||
|
# targetSuffices['q'] = f"q{q}"
|
||||||
|
|
||||||
|
chainVariant = '-'.join([fy['variant'] for fy in chainIteration])
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"\nRunning job {jobIndex} file={sourcePath} variant={chainVariant}")
|
||||||
|
jobIndex += 1
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"label={label if label else 'Falsy'}")
|
||||||
|
ctx.obj['logger'].debug(f"sourceFileBasename={sourceFileBasename}")
|
||||||
|
|
||||||
|
# targetFileBasename = mediaFileProperties.assembleTargetFileBasename(label,
|
||||||
|
# q if len(q_list) > 1 else -1,
|
||||||
|
#
|
||||||
|
targetFileBasename = sourceFileBasename if context['use_tmdb'] and not label else label
|
||||||
|
|
||||||
|
|
||||||
|
targetFilenameTokens = [targetFileBasename]
|
||||||
|
|
||||||
|
if 'se' in targetSuffices.keys():
|
||||||
|
targetFilenameTokens += [targetSuffices['se']]
|
||||||
|
|
||||||
|
# if 'q' in targetSuffices.keys():
|
||||||
|
# targetFilenameTokens += [targetSuffices['q']]
|
||||||
|
for filterYield in chainIteration:
|
||||||
|
|
||||||
|
# filterIdentifier = filterYield['identifier']
|
||||||
|
# filterParameters = filterYield['parameters']
|
||||||
|
# filterSuffices = filterYield['suffices']
|
||||||
|
|
||||||
|
targetFilenameTokens += filterYield['suffices']
|
||||||
|
|
||||||
|
#TODO #387
|
||||||
|
# targetFilename = ((f"{sourceFileBasename}_q{q}" if len(q_list) > 1 else sourceFileBasename)
|
||||||
|
# if context['use_tmdb'] else targetFileBasename)
|
||||||
|
|
||||||
|
targetFilename = f"{'_'.join(targetFilenameTokens)}.{targetExtension}"
|
||||||
|
|
||||||
|
targetPath = os.path.join(output_directory if output_directory else sourceDirectory, targetFilename)
|
||||||
|
|
||||||
|
#TODO: target extension anpassen
|
||||||
|
ctx.obj['logger'].info(f"Creating file {targetFilename}")
|
||||||
|
|
||||||
|
fc.runJob(sourcePath,
|
||||||
|
targetPath,
|
||||||
|
targetFormat,
|
||||||
|
context['video_encoder'],
|
||||||
|
chainIteration)
|
||||||
|
|
||||||
|
#TODO: click.confirm('Warning! This file is not compliant to the defined source schema! Do you want to continue?', abort=True)
|
||||||
|
|
||||||
|
endTime = time.perf_counter()
|
||||||
|
ctx.obj['logger'].info(f"\nDONE\nTime elapsed {endTime - startTime}")
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
ffx()
|
||||||
38
src/ffx/ffx_app.py
Normal file
38
src/ffx/ffx_app.py
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
from textual.app import App
|
||||||
|
|
||||||
|
from .shows_screen import ShowsScreen
|
||||||
|
from .media_details_screen import MediaDetailsScreen
|
||||||
|
|
||||||
|
|
||||||
|
class FfxApp(App):
|
||||||
|
|
||||||
|
TITLE = "FFX"
|
||||||
|
|
||||||
|
BINDINGS = [
|
||||||
|
("q", "quit()", "Quit"),
|
||||||
|
("h", "switch_mode('help')", "Help"),
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = {}):
|
||||||
|
super().__init__()
|
||||||
|
|
||||||
|
# Data 'input' variable
|
||||||
|
self.context = context
|
||||||
|
|
||||||
|
|
||||||
|
def on_mount(self) -> None:
|
||||||
|
|
||||||
|
if 'command' in self.context.keys():
|
||||||
|
|
||||||
|
if self.context['command'] == 'shows':
|
||||||
|
self.push_screen(ShowsScreen())
|
||||||
|
|
||||||
|
if self.context['command'] == 'inspect':
|
||||||
|
self.push_screen(MediaDetailsScreen())
|
||||||
|
|
||||||
|
|
||||||
|
def getContext(self):
|
||||||
|
"""Data 'output' method"""
|
||||||
|
return self.context
|
||||||
|
|
||||||
361
src/ffx/ffx_controller.py
Normal file
361
src/ffx/ffx_controller.py
Normal file
@@ -0,0 +1,361 @@
|
|||||||
|
import os, click
|
||||||
|
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.audio_layout import AudioLayout
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
from ffx.video_encoder import VideoEncoder
|
||||||
|
from ffx.process import executeProcess
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from ffx.track_codec import TrackCodec
|
||||||
|
|
||||||
|
from ffx.constants import DEFAULT_CROP_START, DEFAULT_CROP_LENGTH
|
||||||
|
|
||||||
|
from ffx.filter.quality_filter import QualityFilter
|
||||||
|
from ffx.filter.preset_filter import PresetFilter
|
||||||
|
|
||||||
|
|
||||||
|
class FfxController():
|
||||||
|
|
||||||
|
COMMAND_TOKENS = ['ffmpeg', '-y']
|
||||||
|
NULL_TOKENS = ['-f', 'null', '/dev/null'] # -f null /dev/null
|
||||||
|
|
||||||
|
TEMP_FILE_NAME = "ffmpeg2pass-0.log"
|
||||||
|
|
||||||
|
DEFAULT_VIDEO_ENCODER = VideoEncoder.VP9.label()
|
||||||
|
|
||||||
|
DEFAULT_FILE_FORMAT = 'webm'
|
||||||
|
DEFAULT_FILE_EXTENSION = 'webm'
|
||||||
|
|
||||||
|
INPUT_FILE_EXTENSIONS = ['mkv', 'mp4', 'avi', 'flv', 'webm']
|
||||||
|
|
||||||
|
CHANNEL_MAP_5_1 = 'FL-FL|FR-FR|FC-FC|LFE-LFE|SL-BL|SR-BR:5.1'
|
||||||
|
|
||||||
|
#!
|
||||||
|
SIGNATURE_TAGS = {'RECODED_WITH': 'FFX'}
|
||||||
|
|
||||||
|
def __init__(self,
|
||||||
|
context : dict,
|
||||||
|
targetMediaDescriptor : MediaDescriptor,
|
||||||
|
sourceMediaDescriptor : MediaDescriptor = None):
|
||||||
|
|
||||||
|
self.__context = context
|
||||||
|
self.__sourceMediaDescriptor = sourceMediaDescriptor
|
||||||
|
self.__targetMediaDescriptor = targetMediaDescriptor
|
||||||
|
|
||||||
|
self.__configurationData = self.__context['config'].getData()
|
||||||
|
|
||||||
|
self.__logger = context['logger']
|
||||||
|
|
||||||
|
|
||||||
|
def generateAV1Tokens(self, quality, preset, subIndex : int = 0):
|
||||||
|
|
||||||
|
return [f"-c:v:{int(subIndex)}", 'libsvtav1',
|
||||||
|
'-svtav1-params', f"crf={quality}:preset={preset}:tune=0:enable-overlays=1:scd=1:scm=0",
|
||||||
|
'-pix_fmt', 'yuv420p10le']
|
||||||
|
|
||||||
|
|
||||||
|
# -c:v:0 libvpx-vp9 -row-mt 1 -crf 32 -pass 1 -speed 4 -frame-parallel 0 -g 9999 -aq-mode 0
|
||||||
|
def generateVP9Pass1Tokens(self, quality, subIndex : int = 0):
|
||||||
|
|
||||||
|
return [f"-c:v:{int(subIndex)}",
|
||||||
|
'libvpx-vp9',
|
||||||
|
'-row-mt', '1',
|
||||||
|
'-crf', str(quality),
|
||||||
|
'-pass', '1',
|
||||||
|
'-speed', '4',
|
||||||
|
'-frame-parallel', '0',
|
||||||
|
'-g', '9999',
|
||||||
|
'-aq-mode', '0']
|
||||||
|
|
||||||
|
# -c:v:0 libvpx-vp9 -row-mt 1 -crf 32 -pass 2 -frame-parallel 0 -g 9999 -aq-mode 0 -auto-alt-ref 1 -lag-in-frames 25
|
||||||
|
def generateVP9Pass2Tokens(self, quality, subIndex : int = 0):
|
||||||
|
|
||||||
|
return [f"-c:v:{int(subIndex)}",
|
||||||
|
'libvpx-vp9',
|
||||||
|
'-row-mt', '1',
|
||||||
|
'-crf', str(quality),
|
||||||
|
'-pass', '2',
|
||||||
|
'-frame-parallel', '0',
|
||||||
|
'-g', '9999',
|
||||||
|
'-aq-mode', '0',
|
||||||
|
'-auto-alt-ref', '1',
|
||||||
|
'-lag-in-frames', '25']
|
||||||
|
|
||||||
|
|
||||||
|
def generateCropTokens(self):
|
||||||
|
|
||||||
|
if 'crop_start' in self.__context.keys() and 'crop_length' in self.__context.keys():
|
||||||
|
cropStart = int(self.__context['crop_start'])
|
||||||
|
cropLength = int(self.__context['crop_length'])
|
||||||
|
else:
|
||||||
|
cropStart = DEFAULT_CROP_START
|
||||||
|
cropLength = DEFAULT_CROP_LENGTH
|
||||||
|
|
||||||
|
return ['-ss', str(cropStart), '-t', str(cropLength)]
|
||||||
|
|
||||||
|
|
||||||
|
def generateOutputTokens(self, filePathBase, format = '', ext = ''):
|
||||||
|
outputFilePath = f"{filePathBase}{'.'+str(ext) if ext else ''}"
|
||||||
|
if format:
|
||||||
|
return ['-f', format, outputFilePath]
|
||||||
|
else:
|
||||||
|
return [outputFilePath]
|
||||||
|
|
||||||
|
|
||||||
|
def generateAudioEncodingTokens(self):
|
||||||
|
"""Generates ffmpeg options audio streams including channel remapping, codec and bitrate"""
|
||||||
|
|
||||||
|
audioTokens = []
|
||||||
|
|
||||||
|
targetAudioTrackDescriptors = [td for td in self.__targetMediaDescriptor.getAllTrackDescriptors() if td.getType() == TrackType.AUDIO]
|
||||||
|
|
||||||
|
trackSubIndex = 0
|
||||||
|
for trackDescriptor in targetAudioTrackDescriptors:
|
||||||
|
|
||||||
|
trackAudioLayout = trackDescriptor.getAudioLayout()
|
||||||
|
|
||||||
|
if trackAudioLayout == AudioLayout.LAYOUT_6_1:
|
||||||
|
audioTokens += [f"-c:a:{trackSubIndex}",
|
||||||
|
'libopus',
|
||||||
|
f"-filter:a:{trackSubIndex}",
|
||||||
|
'channelmap=channel_layout=6.1',
|
||||||
|
f"-b:a:{trackSubIndex}",
|
||||||
|
self.__context['bitrates']['dts']]
|
||||||
|
|
||||||
|
if trackAudioLayout == AudioLayout.LAYOUT_5_1:
|
||||||
|
audioTokens += [f"-c:a:{trackSubIndex}",
|
||||||
|
'libopus',
|
||||||
|
f"-filter:a:{trackSubIndex}",
|
||||||
|
f"channelmap={FfxController.CHANNEL_MAP_5_1}",
|
||||||
|
f"-b:a:{trackSubIndex}",
|
||||||
|
self.__context['bitrates']['ac3']]
|
||||||
|
|
||||||
|
if trackAudioLayout == AudioLayout.LAYOUT_STEREO:
|
||||||
|
audioTokens += [f"-c:a:{trackSubIndex}",
|
||||||
|
'libopus',
|
||||||
|
f"-b:a:{trackSubIndex}",
|
||||||
|
self.__context['bitrates']['stereo']]
|
||||||
|
|
||||||
|
if trackAudioLayout == AudioLayout.LAYOUT_6CH:
|
||||||
|
audioTokens += [f"-c:a:{trackSubIndex}",
|
||||||
|
'libopus',
|
||||||
|
f"-filter:a:{trackSubIndex}",
|
||||||
|
f"channelmap={FfxController.CHANNEL_MAP_5_1}",
|
||||||
|
f"-b:a:{trackSubIndex}",
|
||||||
|
self.__context['bitrates']['ac3']]
|
||||||
|
trackSubIndex += 1
|
||||||
|
return audioTokens
|
||||||
|
|
||||||
|
|
||||||
|
# -disposition:s:0 default -disposition:s:1 0
|
||||||
|
def generateDispositionTokens(self):
|
||||||
|
|
||||||
|
targetTrackDescriptors = self.__targetMediaDescriptor.getAllTrackDescriptors()
|
||||||
|
|
||||||
|
sourceTrackDescriptors = ([] if self.__sourceMediaDescriptor is None
|
||||||
|
else self.__sourceMediaDescriptor.getAllTrackDescriptors())
|
||||||
|
|
||||||
|
dispositionTokens = []
|
||||||
|
|
||||||
|
for trackIndex in range(len(targetTrackDescriptors)):
|
||||||
|
|
||||||
|
td = targetTrackDescriptors[trackIndex]
|
||||||
|
|
||||||
|
#HINT: No dispositions for pgs subtitle tracks that have no external file source
|
||||||
|
if (td.getExternalSourceFilePath()
|
||||||
|
or td.getCodec() != TrackCodec.PGS):
|
||||||
|
|
||||||
|
subIndex = td.getSubIndex()
|
||||||
|
streamIndicator = td.getType().indicator()
|
||||||
|
|
||||||
|
|
||||||
|
sourceDispositionSet = sourceTrackDescriptors[td.getSourceIndex()].getDispositionSet() if sourceTrackDescriptors else set()
|
||||||
|
|
||||||
|
#TODO: Alles discarden was im targetDescriptor vorhanden ist (?)
|
||||||
|
sourceDispositionSet.discard(TrackDisposition.DEFAULT)
|
||||||
|
|
||||||
|
dispositionSet = td.getDispositionSet() | sourceDispositionSet
|
||||||
|
|
||||||
|
if dispositionSet:
|
||||||
|
dispositionTokens += [f"-disposition:{streamIndicator}:{subIndex}", '+'.join([d.label() for d in dispositionSet])]
|
||||||
|
else:
|
||||||
|
dispositionTokens += [f"-disposition:{streamIndicator}:{subIndex}", '0']
|
||||||
|
|
||||||
|
return dispositionTokens
|
||||||
|
|
||||||
|
|
||||||
|
def generateMetadataTokens(self):
|
||||||
|
|
||||||
|
metadataTokens = []
|
||||||
|
|
||||||
|
metadataConfiguration = self.__configurationData['metadata'] if 'metadata' in self.__configurationData.keys() else {}
|
||||||
|
|
||||||
|
signatureTags = metadataConfiguration['signature'] if 'signature' in metadataConfiguration.keys() else {}
|
||||||
|
removeGlobalKeys = metadataConfiguration['remove'] if 'remove' in metadataConfiguration.keys() else []
|
||||||
|
removeTrackKeys = metadataConfiguration['streams']['remove'] if 'streams' in metadataConfiguration.keys() and 'remove' in metadataConfiguration['streams'].keys() else []
|
||||||
|
|
||||||
|
mediaTags = {k:v for k,v in self.__targetMediaDescriptor.getTags().items() if not k in removeGlobalKeys}
|
||||||
|
|
||||||
|
if (not 'no_signature' in self.__context.keys()
|
||||||
|
or not self.__context['no_signature']):
|
||||||
|
outputMediaTags = mediaTags | signatureTags
|
||||||
|
else:
|
||||||
|
outputMediaTags = mediaTags
|
||||||
|
|
||||||
|
for tagKey, tagValue in outputMediaTags.items():
|
||||||
|
metadataTokens += [f"-metadata:g",
|
||||||
|
f"{tagKey}={tagValue}"]
|
||||||
|
|
||||||
|
for removeKey in removeGlobalKeys:
|
||||||
|
metadataTokens += [f"-metadata:g",
|
||||||
|
f"{removeKey}="]
|
||||||
|
|
||||||
|
|
||||||
|
removeMkvmergeMetadata = (not 'keep_mkvmerge_metadata' in self.__context.keys()
|
||||||
|
or not self.__context['keep_mkvmerge_metadata'])
|
||||||
|
|
||||||
|
#HINT: With current ffmpeg version track metadata tags are not passed to the outfile
|
||||||
|
for td in self.__targetMediaDescriptor.getAllTrackDescriptors():
|
||||||
|
|
||||||
|
typeIndicator = td.getType().indicator()
|
||||||
|
subIndex = td.getSubIndex()
|
||||||
|
|
||||||
|
for tagKey, tagValue in td.getTags().items():
|
||||||
|
|
||||||
|
if not tagKey in removeTrackKeys:
|
||||||
|
metadataTokens += [f"-metadata:s:{typeIndicator}:{subIndex}",
|
||||||
|
f"{tagKey}={tagValue}"]
|
||||||
|
|
||||||
|
for removeKey in removeTrackKeys:
|
||||||
|
metadataTokens += [f"-metadata:s:{typeIndicator}:{subIndex}",
|
||||||
|
f"{removeKey}="]
|
||||||
|
|
||||||
|
|
||||||
|
return metadataTokens
|
||||||
|
|
||||||
|
|
||||||
|
def runJob(self,
|
||||||
|
sourcePath,
|
||||||
|
targetPath,
|
||||||
|
targetFormat: str = '',
|
||||||
|
videoEncoder: VideoEncoder = VideoEncoder.VP9,
|
||||||
|
chainIteration: list = []):
|
||||||
|
# quality: int = DEFAULT_QUALITY,
|
||||||
|
# preset: int = DEFAULT_AV1_PRESET):
|
||||||
|
|
||||||
|
qualityFilters = [fy for fy in chainIteration if fy['identifier'] == 'quality']
|
||||||
|
presetFilters = [fy for fy in chainIteration if fy['identifier'] == 'preset']
|
||||||
|
denoiseFilters = [fy for fy in chainIteration if fy['identifier'] == 'nlmeans']
|
||||||
|
|
||||||
|
quality = qualityFilters[0]['parameters']['quality'] if qualityFilters else QualityFilter.DEFAULT_QUALITY
|
||||||
|
preset = presetFilters[0]['parameters']['preset'] if presetFilters else PresetFilter.DEFAULT_PRESET
|
||||||
|
|
||||||
|
|
||||||
|
denoiseTokens = denoiseFilters[0]['tokens'] if denoiseFilters else []
|
||||||
|
|
||||||
|
|
||||||
|
commandTokens = FfxController.COMMAND_TOKENS + ['-i', sourcePath]
|
||||||
|
|
||||||
|
if videoEncoder == VideoEncoder.AV1:
|
||||||
|
|
||||||
|
commandSequence = (commandTokens
|
||||||
|
+ self.__targetMediaDescriptor.getImportFileTokens()
|
||||||
|
+ self.__targetMediaDescriptor.getInputMappingTokens()
|
||||||
|
+ self.generateDispositionTokens())
|
||||||
|
|
||||||
|
# Optional tokens
|
||||||
|
commandSequence += self.generateMetadataTokens()
|
||||||
|
commandSequence += denoiseTokens
|
||||||
|
|
||||||
|
commandSequence += (self.generateAudioEncodingTokens()
|
||||||
|
+ self.generateAV1Tokens(int(quality), int(preset))
|
||||||
|
+ self.generateAudioEncodingTokens())
|
||||||
|
|
||||||
|
if self.__context['perform_crop']:
|
||||||
|
commandSequence += FfxController.generateCropTokens()
|
||||||
|
|
||||||
|
commandSequence += self.generateOutputTokens(targetPath,
|
||||||
|
targetFormat)
|
||||||
|
|
||||||
|
self.__logger.debug(f"FfxController.runJob(): Running command sequence")
|
||||||
|
|
||||||
|
if not self.__context['dry_run']:
|
||||||
|
executeProcess(commandSequence, context = self.__context)
|
||||||
|
|
||||||
|
|
||||||
|
if videoEncoder == VideoEncoder.VP9:
|
||||||
|
|
||||||
|
commandSequence1 = (commandTokens
|
||||||
|
+ self.__targetMediaDescriptor.getInputMappingTokens(only_video=True))
|
||||||
|
|
||||||
|
# Optional tokens
|
||||||
|
#NOTE: Filters and so needs to run on the first pass as well, as here
|
||||||
|
# the required bitrate for the second run is determined and recorded
|
||||||
|
# TODO: Results seems to be slightly better with first pass omitted,
|
||||||
|
# Confirm or find better filter settings for 2-pass
|
||||||
|
# commandSequence1 += self.__context['denoiser'].generateDenoiseTokens()
|
||||||
|
|
||||||
|
commandSequence1 += self.generateVP9Pass1Tokens(int(quality))
|
||||||
|
|
||||||
|
if self.__context['perform_crop']:
|
||||||
|
commandSequence1 += self.generateCropTokens()
|
||||||
|
|
||||||
|
commandSequence1 += FfxController.NULL_TOKENS
|
||||||
|
|
||||||
|
if os.path.exists(FfxController.TEMP_FILE_NAME):
|
||||||
|
os.remove(FfxController.TEMP_FILE_NAME)
|
||||||
|
|
||||||
|
self.__logger.debug(f"FfxController.runJob(): Running command sequence 1")
|
||||||
|
|
||||||
|
if not self.__context['dry_run']:
|
||||||
|
executeProcess(commandSequence1, context = self.__context)
|
||||||
|
|
||||||
|
commandSequence2 = (commandTokens
|
||||||
|
+ self.__targetMediaDescriptor.getImportFileTokens()
|
||||||
|
+ self.__targetMediaDescriptor.getInputMappingTokens()
|
||||||
|
+ self.generateDispositionTokens())
|
||||||
|
|
||||||
|
# Optional tokens
|
||||||
|
commandSequence2 += self.generateMetadataTokens()
|
||||||
|
commandSequence2 += denoiseTokens
|
||||||
|
|
||||||
|
commandSequence2 += self.generateVP9Pass2Tokens(int(quality)) + self.generateAudioEncodingTokens()
|
||||||
|
|
||||||
|
if self.__context['perform_crop']:
|
||||||
|
commandSequence2 += self.generateCropTokens()
|
||||||
|
|
||||||
|
commandSequence2 += self.generateOutputTokens(targetPath,
|
||||||
|
targetFormat)
|
||||||
|
|
||||||
|
self.__logger.debug(f"FfxController.runJob(): Running command sequence 2")
|
||||||
|
|
||||||
|
if not self.__context['dry_run']:
|
||||||
|
out, err, rc = executeProcess(commandSequence2, context = self.__context)
|
||||||
|
if rc:
|
||||||
|
raise click.ClickException(f"Command resulted in error: rc={rc} error={err}")
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def createEmptyFile(self,
|
||||||
|
path: str = 'empty.mkv',
|
||||||
|
sizeX: int = 1280,
|
||||||
|
sizeY: int = 720,
|
||||||
|
rate: int = 25,
|
||||||
|
length: int = 10):
|
||||||
|
|
||||||
|
commandTokens = FfxController.COMMAND_TOKENS
|
||||||
|
|
||||||
|
commandTokens += ['-f',
|
||||||
|
'lavfi',
|
||||||
|
'-i',
|
||||||
|
f"color=size={sizeX}x{sizeY}:rate={rate}:color=black",
|
||||||
|
'-f',
|
||||||
|
'lavfi',
|
||||||
|
'-i',
|
||||||
|
'anullsrc=channel_layout=stereo:sample_rate=44100',
|
||||||
|
'-t',
|
||||||
|
str(length),
|
||||||
|
path]
|
||||||
|
|
||||||
|
out, err, rc = executeProcess(commandTokens, context = self.__context)
|
||||||
124
src/ffx/ffx_tests.py
Executable file
124
src/ffx/ffx_tests.py
Executable file
@@ -0,0 +1,124 @@
|
|||||||
|
#! /usr/bin/python3
|
||||||
|
|
||||||
|
import os, logging, click
|
||||||
|
|
||||||
|
from ffx.configuration_controller import ConfigurationController
|
||||||
|
|
||||||
|
from ffx.file_properties import FileProperties
|
||||||
|
from ffx.ffx_controller import FfxController
|
||||||
|
|
||||||
|
from ffx.test.helper import createMediaTestFile
|
||||||
|
|
||||||
|
from ffx.test.scenario import Scenario
|
||||||
|
from ffx.tmdb_controller import TMDB_API_KEY_NOT_PRESENT_EXCEPTION
|
||||||
|
|
||||||
|
|
||||||
|
@click.group()
|
||||||
|
@click.pass_context
|
||||||
|
@click.option('-v', '--verbose', type=int, default=0, help='Set verbosity of output')
|
||||||
|
@click.option("--dry-run", is_flag=True, default=False)
|
||||||
|
def ffx(ctx, verbose, dry_run):
|
||||||
|
"""FFX"""
|
||||||
|
|
||||||
|
ctx.obj = {}
|
||||||
|
|
||||||
|
ctx.obj['config'] = ConfigurationController()
|
||||||
|
|
||||||
|
ctx.obj['database'] = None
|
||||||
|
ctx.obj['dry_run'] = dry_run
|
||||||
|
|
||||||
|
ctx.obj['verbosity'] = verbose
|
||||||
|
|
||||||
|
# Critical 50
|
||||||
|
# Error 40
|
||||||
|
# Warning 30
|
||||||
|
# Info 20
|
||||||
|
# Debug 10
|
||||||
|
fileLogVerbosity = max(40 - verbose * 10, 10)
|
||||||
|
consoleLogVerbosity = max(20 - verbose * 10, 10)
|
||||||
|
|
||||||
|
ctx.obj['logger'] = logging.getLogger('FFX Tests')
|
||||||
|
ctx.obj['logger'].setLevel(logging.DEBUG)
|
||||||
|
|
||||||
|
ctx.obj['report_logger'] = logging.getLogger('FFX Test Result')
|
||||||
|
ctx.obj['report_logger'].setLevel(logging.INFO)
|
||||||
|
|
||||||
|
ffxFileHandler = logging.FileHandler(ctx.obj['config'].getLogFilePath())
|
||||||
|
ffxFileHandler.setLevel(fileLogVerbosity)
|
||||||
|
ffxConsoleHandler = logging.StreamHandler()
|
||||||
|
ffxConsoleHandler.setLevel(consoleLogVerbosity)
|
||||||
|
|
||||||
|
if os.path.isfile('ffx_test_report.log'):
|
||||||
|
os.unlink('ffx_test_report.log')
|
||||||
|
ffxTestReportFileHandler = logging.FileHandler('ffx_test_report.log')
|
||||||
|
|
||||||
|
fileFormatter = logging.Formatter(
|
||||||
|
'%(asctime)s - %(name)s - %(levelname)s - %(message)s')
|
||||||
|
ffxFileHandler.setFormatter(fileFormatter)
|
||||||
|
consoleFormatter = logging.Formatter(
|
||||||
|
'%(message)s')
|
||||||
|
ffxConsoleHandler.setFormatter(consoleFormatter)
|
||||||
|
reportFormatter = logging.Formatter(
|
||||||
|
'%(message)s')
|
||||||
|
ffxTestReportFileHandler.setFormatter(reportFormatter)
|
||||||
|
|
||||||
|
ctx.obj['logger'].addHandler(ffxConsoleHandler)
|
||||||
|
ctx.obj['logger'].addHandler(ffxFileHandler)
|
||||||
|
|
||||||
|
ctx.obj['report_logger'].addHandler(ffxConsoleHandler)
|
||||||
|
ctx.obj['report_logger'].addHandler(ffxTestReportFileHandler)
|
||||||
|
|
||||||
|
|
||||||
|
# Another subcommand
|
||||||
|
@ffx.command()
|
||||||
|
@click.pass_context
|
||||||
|
@click.option('--scenario', type=str, default='', help='Only run tests from this scenario')
|
||||||
|
@click.option('--variant', type=str, default='', help='Only run variants beginning like this')
|
||||||
|
@click.option('--limit', type=int, default=0, help='Only run this number of tests')
|
||||||
|
def run(ctx, scenario, variant, limit):
|
||||||
|
"""Run ffx test sequences"""
|
||||||
|
|
||||||
|
ctx.obj['logger'].info('Starting FFX test runs')
|
||||||
|
ctx.obj['test_passed_counter'] = 0
|
||||||
|
ctx.obj['test_failed_counter'] = 0
|
||||||
|
|
||||||
|
ctx.obj['test_variant'] = variant
|
||||||
|
ctx.obj['test_limit'] = limit
|
||||||
|
|
||||||
|
for si in Scenario.list():
|
||||||
|
|
||||||
|
try:
|
||||||
|
SCEN = Scenario.getClassReference(si)
|
||||||
|
scen = SCEN(ctx.obj)
|
||||||
|
|
||||||
|
if scenario and scenario != scen.getScenario():
|
||||||
|
continue
|
||||||
|
|
||||||
|
ctx.obj['logger'].debug(f"Running scenario {si}")
|
||||||
|
|
||||||
|
scen.run()
|
||||||
|
|
||||||
|
except TMDB_API_KEY_NOT_PRESENT_EXCEPTION:
|
||||||
|
ctx.obj['logger'].info(f"TMDB_API_KEY not set: Skipping {SCEN.__class__.__name__}")
|
||||||
|
|
||||||
|
ctx.obj['logger'].info(f"\n{ctx.obj['test_passed_counter']} tests passed")
|
||||||
|
ctx.obj['logger'].info(f"{ctx.obj['test_failed_counter']} test failed")
|
||||||
|
ctx.obj['logger'].info('\nDone.')
|
||||||
|
|
||||||
|
|
||||||
|
@ffx.command()
|
||||||
|
@click.pass_context
|
||||||
|
@click.argument('paths', nargs=-1)
|
||||||
|
def dupe(ctx, paths):
|
||||||
|
|
||||||
|
existingSourcePaths = [p for p in paths if os.path.isfile(p) and p.split('.')[-1] in FfxController.INPUT_FILE_EXTENSIONS]
|
||||||
|
|
||||||
|
for sourcePath in existingSourcePaths:
|
||||||
|
|
||||||
|
sourceFileProperties = FileProperties(ctx.obj, sourcePath)
|
||||||
|
sourceMediaDescriptor = sourceFileProperties.getMediaDescriptor()
|
||||||
|
|
||||||
|
createMediaTestFile(sourceMediaDescriptor, baseName='dupe')
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
ffx()
|
||||||
201
src/ffx/file_properties.py
Normal file
201
src/ffx/file_properties.py
Normal file
@@ -0,0 +1,201 @@
|
|||||||
|
import os, re, json
|
||||||
|
|
||||||
|
from .media_descriptor import MediaDescriptor
|
||||||
|
from .pattern_controller import PatternController
|
||||||
|
|
||||||
|
from .process import executeProcess
|
||||||
|
|
||||||
|
from ffx.model.pattern import Pattern
|
||||||
|
|
||||||
|
|
||||||
|
class FileProperties():
|
||||||
|
|
||||||
|
FILE_EXTENSIONS = ['mkv', 'mp4', 'avi', 'flv', 'webm']
|
||||||
|
|
||||||
|
SE_INDICATOR_PATTERN = '([sS][0-9]+[eE][0-9]+)'
|
||||||
|
SEASON_EPISODE_INDICATOR_MATCH = '[sS]([0-9]+)[eE]([0-9]+)'
|
||||||
|
EPISODE_INDICATOR_MATCH = '[eE]([0-9]+)'
|
||||||
|
|
||||||
|
DEFAULT_INDEX_DIGITS = 3
|
||||||
|
|
||||||
|
def __init__(self, context, sourcePath):
|
||||||
|
|
||||||
|
self.context = context
|
||||||
|
|
||||||
|
self.__logger = context['logger']
|
||||||
|
|
||||||
|
# Separate basedir, basename and extension for current source file
|
||||||
|
self.__sourcePath = sourcePath
|
||||||
|
|
||||||
|
self.__sourceDirectory = os.path.dirname(self.__sourcePath)
|
||||||
|
self.__sourceFilename = os.path.basename(self.__sourcePath)
|
||||||
|
|
||||||
|
sourcePathTokens = self.__sourceFilename.split('.')
|
||||||
|
|
||||||
|
if sourcePathTokens[-1] in FileProperties.FILE_EXTENSIONS:
|
||||||
|
self.__sourceFileBasename = '.'.join(sourcePathTokens[:-1])
|
||||||
|
self.__sourceFilenameExtension = sourcePathTokens[-1]
|
||||||
|
else:
|
||||||
|
self.__sourceFileBasename = self.__sourceFilename
|
||||||
|
self.__sourceFilenameExtension = ''
|
||||||
|
|
||||||
|
self.__pc = PatternController(context)
|
||||||
|
|
||||||
|
# Checking if database contains matching pattern
|
||||||
|
matchResult = self.__pc.matchFilename(self.__sourceFilename)
|
||||||
|
|
||||||
|
self.__logger.debug(f"FileProperties.__init__(): Match result: {matchResult}")
|
||||||
|
|
||||||
|
self.__pattern: Pattern = matchResult['pattern'] if matchResult else None
|
||||||
|
|
||||||
|
if matchResult:
|
||||||
|
databaseMatchedGroups = matchResult['match'].groups()
|
||||||
|
self.__logger.debug(f"FileProperties.__init__(): Matched groups: {databaseMatchedGroups}")
|
||||||
|
|
||||||
|
seIndicator = databaseMatchedGroups[0]
|
||||||
|
|
||||||
|
se_match = re.search(FileProperties.SEASON_EPISODE_INDICATOR_MATCH, seIndicator)
|
||||||
|
e_match = re.search(FileProperties.EPISODE_INDICATOR_MATCH, seIndicator)
|
||||||
|
|
||||||
|
else:
|
||||||
|
self.__logger.debug(f"FileProperties.__init__(): Checking file name for indicator {self.__sourceFilename}")
|
||||||
|
|
||||||
|
se_match = re.search(FileProperties.SEASON_EPISODE_INDICATOR_MATCH, self.__sourceFilename)
|
||||||
|
e_match = re.search(FileProperties.EPISODE_INDICATOR_MATCH, self.__sourceFilename)
|
||||||
|
|
||||||
|
if se_match is not None:
|
||||||
|
self.__season = int(se_match.group(1))
|
||||||
|
self.__episode = int(se_match.group(2))
|
||||||
|
elif e_match is not None:
|
||||||
|
self.__season = -1
|
||||||
|
self.__episode = int(e_match.group(1))
|
||||||
|
else:
|
||||||
|
self.__season = -1
|
||||||
|
self.__episode = -1
|
||||||
|
|
||||||
|
|
||||||
|
def getFormatData(self):
|
||||||
|
"""
|
||||||
|
"format": {
|
||||||
|
"filename": "Downloads/nagatoro_s02/nagatoro_s01e02.mkv",
|
||||||
|
"nb_streams": 18,
|
||||||
|
"nb_programs": 0,
|
||||||
|
"nb_stream_groups": 0,
|
||||||
|
"format_name": "matroska,webm",
|
||||||
|
"format_long_name": "Matroska / WebM",
|
||||||
|
"start_time": "0.000000",
|
||||||
|
"duration": "1420.063000",
|
||||||
|
"size": "1489169824",
|
||||||
|
"bit_rate": "8389316",
|
||||||
|
"probe_score": 100,
|
||||||
|
"tags": {
|
||||||
|
"PUBLISHER": "Crunchyroll",
|
||||||
|
"ENCODER": "Lavf58.29.100"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
"""
|
||||||
|
|
||||||
|
# ffprobe -hide_banner -show_format -of json
|
||||||
|
ffprobeOutput, ffprobeError, returnCode = executeProcess(["ffprobe",
|
||||||
|
"-hide_banner",
|
||||||
|
"-show_format",
|
||||||
|
"-of", "json",
|
||||||
|
self.__sourcePath]) #,
|
||||||
|
#context = self.context)
|
||||||
|
|
||||||
|
if 'Invalid data found when processing input' in ffprobeError:
|
||||||
|
raise Exception(f"File {self.__sourcePath} does not contain valid stream data")
|
||||||
|
|
||||||
|
if returnCode != 0:
|
||||||
|
raise Exception(f"ffprobe returned with error {returnCode}")
|
||||||
|
|
||||||
|
return json.loads(ffprobeOutput)['format']
|
||||||
|
|
||||||
|
|
||||||
|
def getStreamData(self):
|
||||||
|
"""Returns ffprobe stream data as array with elements according to the following example
|
||||||
|
{
|
||||||
|
"index": 4,
|
||||||
|
"codec_name": "hdmv_pgs_subtitle",
|
||||||
|
"codec_long_name": "HDMV Presentation Graphic Stream subtitles",
|
||||||
|
"codec_type": "subtitle",
|
||||||
|
"codec_tag_string": "[0][0][0][0]",
|
||||||
|
"codec_tag": "0x0000",
|
||||||
|
"r_frame_rate": "0/0",
|
||||||
|
"avg_frame_rate": "0/0",
|
||||||
|
"time_base": "1/1000",
|
||||||
|
"start_pts": 0,
|
||||||
|
"start_time": "0.000000",
|
||||||
|
"duration_ts": 1421035,
|
||||||
|
"duration": "1421.035000",
|
||||||
|
"disposition": {
|
||||||
|
"default": 1,
|
||||||
|
"dub": 0,
|
||||||
|
"original": 0,
|
||||||
|
"comment": 0,
|
||||||
|
"lyrics": 0,
|
||||||
|
"karaoke": 0,
|
||||||
|
"forced": 0,
|
||||||
|
"hearing_impaired": 0,
|
||||||
|
"visual_impaired": 0,
|
||||||
|
"clean_effects": 0,
|
||||||
|
"attached_pic": 0,
|
||||||
|
"timed_thumbnails": 0,
|
||||||
|
"non_diegetic": 0,
|
||||||
|
"captions": 0,
|
||||||
|
"descriptions": 0,
|
||||||
|
"metadata": 0,
|
||||||
|
"dependent": 0,
|
||||||
|
"still_image": 0
|
||||||
|
},
|
||||||
|
"tags": {
|
||||||
|
"language": "ger",
|
||||||
|
"title": "German Full"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
"""
|
||||||
|
|
||||||
|
# ffprobe -hide_banner -show_streams -of json
|
||||||
|
ffprobeOutput, ffprobeError, returnCode = executeProcess(["ffprobe",
|
||||||
|
"-hide_banner",
|
||||||
|
"-show_streams",
|
||||||
|
"-of", "json",
|
||||||
|
self.__sourcePath]) #,
|
||||||
|
#context = self.context)
|
||||||
|
|
||||||
|
if 'Invalid data found when processing input' in ffprobeError:
|
||||||
|
raise Exception(f"File {self.__sourcePath} does not contain valid stream data")
|
||||||
|
|
||||||
|
|
||||||
|
if returnCode != 0:
|
||||||
|
raise Exception(f"ffprobe returned with error {returnCode}")
|
||||||
|
|
||||||
|
|
||||||
|
return json.loads(ffprobeOutput)['streams']
|
||||||
|
|
||||||
|
|
||||||
|
def getMediaDescriptor(self):
|
||||||
|
return MediaDescriptor.fromFfprobe(self.context, self.getFormatData(), self.getStreamData())
|
||||||
|
|
||||||
|
|
||||||
|
def getShowId(self) -> int:
|
||||||
|
"""Result is -1 if the filename did not match anything in database"""
|
||||||
|
return self.__pattern.getShowId() if self.__pattern is not None else -1
|
||||||
|
|
||||||
|
def getPattern(self) -> Pattern:
|
||||||
|
"""Result is None if the filename did not match anything in database"""
|
||||||
|
return self.__pattern
|
||||||
|
|
||||||
|
|
||||||
|
def getSeason(self) -> int:
|
||||||
|
return int(self.__season)
|
||||||
|
|
||||||
|
def getEpisode(self) -> int:
|
||||||
|
return int(self.__episode)
|
||||||
|
|
||||||
|
|
||||||
|
def getFilename(self):
|
||||||
|
return self.__sourceFilename
|
||||||
|
|
||||||
|
def getFileBasename(self):
|
||||||
|
return self.__sourceFileBasename
|
||||||
0
src/ffx/filter/__init__.py
Normal file
0
src/ffx/filter/__init__.py
Normal file
17
src/ffx/filter/filter.py
Normal file
17
src/ffx/filter/filter.py
Normal file
@@ -0,0 +1,17 @@
|
|||||||
|
import itertools
|
||||||
|
|
||||||
|
|
||||||
|
class Filter():
|
||||||
|
|
||||||
|
filterChain: list = []
|
||||||
|
|
||||||
|
def __init__(self, filter):
|
||||||
|
|
||||||
|
self.filterChain.append(filter)
|
||||||
|
|
||||||
|
def getFilterChain(self):
|
||||||
|
return self.filterChain
|
||||||
|
|
||||||
|
def getChainYield(self):
|
||||||
|
for fy in itertools.product(*[f.getYield() for f in self.filterChain]):
|
||||||
|
yield fy
|
||||||
162
src/ffx/filter/nlmeans_filter.py
Normal file
162
src/ffx/filter/nlmeans_filter.py
Normal file
@@ -0,0 +1,162 @@
|
|||||||
|
import itertools
|
||||||
|
|
||||||
|
from .filter import Filter
|
||||||
|
|
||||||
|
|
||||||
|
class NlmeansFilter(Filter):
|
||||||
|
|
||||||
|
IDENTIFIER = 'nlmeans'
|
||||||
|
|
||||||
|
DEFAULT_STRENGTH: float = 2.8
|
||||||
|
DEFAULT_PATCH_SIZE: int = 13
|
||||||
|
DEFAULT_CHROMA_PATCH_SIZE: int = 9
|
||||||
|
DEFAULT_RESEARCH_WINDOW: int = 23
|
||||||
|
DEFAULT_CHROMA_RESEARCH_WINDOW: int= 17
|
||||||
|
|
||||||
|
STRENGTH_KEY = 'strength'
|
||||||
|
PATCH_SIZE_KEY = 'patch_size'
|
||||||
|
CHROMA_PATCH_SIZE_KEY = 'chroma_patch_size'
|
||||||
|
RESEARCH_WINDOW_KEY = 'research_window'
|
||||||
|
CHROMA_RESEARCH_WINDOW_KEY = 'chroma_research_window'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, **kwargs):
|
||||||
|
|
||||||
|
self.__useHardware = kwargs.get('use_hardware', False)
|
||||||
|
|
||||||
|
self.__strengthList = []
|
||||||
|
strength = kwargs.get(NlmeansFilter.STRENGTH_KEY, '')
|
||||||
|
if strength:
|
||||||
|
strengthTokens = strength.split(',')
|
||||||
|
for st in strengthTokens:
|
||||||
|
try:
|
||||||
|
strengthValue = float(st)
|
||||||
|
except:
|
||||||
|
raise ValueError('NlmeansFilter: Strength value has to be of type float')
|
||||||
|
if strengthValue < 1.0 or strengthValue > 30.0:
|
||||||
|
raise ValueError('NlmeansFilter: Strength value has to be between 1.0 and 30.0')
|
||||||
|
self.__strengthList.append(strengthValue)
|
||||||
|
else:
|
||||||
|
self.__strengthList = [NlmeansFilter.DEFAULT_STRENGTH]
|
||||||
|
|
||||||
|
self.__patchSizeList = []
|
||||||
|
patchSize = kwargs.get(NlmeansFilter.PATCH_SIZE_KEY, '')
|
||||||
|
if patchSize:
|
||||||
|
patchSizeTokens = patchSize.split(',')
|
||||||
|
for pst in patchSizeTokens:
|
||||||
|
try:
|
||||||
|
patchSizeValue = int(pst)
|
||||||
|
except:
|
||||||
|
raise ValueError('NlmeansFilter: Patch size value has to be of type int')
|
||||||
|
if patchSizeValue < 0 or patchSizeValue > 99:
|
||||||
|
raise ValueError('NlmeansFilter: Patch size value has to be between 0 and 99')
|
||||||
|
if patchSizeValue % 2 == 0:
|
||||||
|
raise ValueError('NlmeansFilter: Patch size value has to an odd number')
|
||||||
|
self.__patchSizeList.append(patchSizeValue)
|
||||||
|
else:
|
||||||
|
self.__patchSizeList = [NlmeansFilter.DEFAULT_PATCH_SIZE]
|
||||||
|
|
||||||
|
self.__chromaPatchSizeList = []
|
||||||
|
chromaPatchSize = kwargs.get(NlmeansFilter.CHROMA_PATCH_SIZE_KEY, '')
|
||||||
|
if chromaPatchSize:
|
||||||
|
chromaPatchSizeTokens = chromaPatchSize.split(',')
|
||||||
|
for cpst in chromaPatchSizeTokens:
|
||||||
|
try:
|
||||||
|
chromaPatchSizeValue = int(pst)
|
||||||
|
except:
|
||||||
|
raise ValueError('NlmeansFilter: Chroma patch size value has to be of type int')
|
||||||
|
if chromaPatchSizeValue < 0 or chromaPatchSizeValue > 99:
|
||||||
|
raise ValueError('NlmeansFilter: Chroma patch value has to be between 0 and 99')
|
||||||
|
if chromaPatchSizeValue % 2 == 0:
|
||||||
|
raise ValueError('NlmeansFilter: Chroma patch value has to an odd number')
|
||||||
|
self.__chromaPatchSizeList.append(chromaPatchSizeValue)
|
||||||
|
else:
|
||||||
|
self.__chromaPatchSizeList = [NlmeansFilter.DEFAULT_CHROMA_PATCH_SIZE]
|
||||||
|
|
||||||
|
self.__researchWindowList = []
|
||||||
|
researchWindow = kwargs.get(NlmeansFilter.RESEARCH_WINDOW_KEY, '')
|
||||||
|
if researchWindow:
|
||||||
|
researchWindowTokens = researchWindow.split(',')
|
||||||
|
for rwt in researchWindowTokens:
|
||||||
|
try:
|
||||||
|
researchWindowValue = int(rwt)
|
||||||
|
except:
|
||||||
|
raise ValueError('NlmeansFilter: Research window value has to be of type int')
|
||||||
|
if researchWindowValue < 0 or researchWindowValue > 99:
|
||||||
|
raise ValueError('NlmeansFilter: Research window value has to be between 0 and 99')
|
||||||
|
if researchWindowValue % 2 == 0:
|
||||||
|
raise ValueError('NlmeansFilter: Research window value has to an odd number')
|
||||||
|
self.__researchWindowList.append(researchWindowValue)
|
||||||
|
else:
|
||||||
|
self.__researchWindowList = [NlmeansFilter.DEFAULT_RESEARCH_WINDOW]
|
||||||
|
|
||||||
|
self.__chromaResearchWindowList = []
|
||||||
|
chromaResearchWindow = kwargs.get(NlmeansFilter.CHROMA_RESEARCH_WINDOW_KEY, '')
|
||||||
|
if chromaResearchWindow:
|
||||||
|
chromaResearchWindowTokens = chromaResearchWindow.split(',')
|
||||||
|
for crwt in chromaResearchWindowTokens:
|
||||||
|
try:
|
||||||
|
chromaResearchWindowValue = int(crwt)
|
||||||
|
except:
|
||||||
|
raise ValueError('NlmeansFilter: Chroma research window value has to be of type int')
|
||||||
|
if chromaResearchWindowValue < 0 or chromaResearchWindowValue > 99:
|
||||||
|
raise ValueError('NlmeansFilter: Chroma research window value has to be between 0 and 99')
|
||||||
|
if chromaResearchWindowValue % 2 == 0:
|
||||||
|
raise ValueError('NlmeansFilter: Chroma research window value has to an odd number')
|
||||||
|
self.__chromaResearchWindowList.append(chromaResearchWindowValue)
|
||||||
|
else:
|
||||||
|
self.__chromaResearchWindowList = [NlmeansFilter.DEFAULT_CHROMA_RESEARCH_WINDOW]
|
||||||
|
|
||||||
|
super().__init__(self)
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self, iteration):
|
||||||
|
|
||||||
|
strength = iteration[0]
|
||||||
|
patchSize = iteration[1]
|
||||||
|
chromaPatchSize = iteration[2]
|
||||||
|
researchWindow = iteration[3]
|
||||||
|
chromaResearchWindow = iteration[4]
|
||||||
|
|
||||||
|
suffices = []
|
||||||
|
|
||||||
|
if len(self.__strengthList) > 1:
|
||||||
|
suffices += [f"ds{strength}"]
|
||||||
|
if len(self.__patchSizeList) > 1:
|
||||||
|
suffices += [f"dp{patchSize}"]
|
||||||
|
if len(self.__chromaPatchSizeList) > 1:
|
||||||
|
suffices += [f"dpc{chromaPatchSize}"]
|
||||||
|
if len(self.__researchWindowList) > 1:
|
||||||
|
suffices += [f"dr{researchWindow}"]
|
||||||
|
if len(self.__chromaResearchWindowList) > 1:
|
||||||
|
suffices += [f"drc{chromaResearchWindow}"]
|
||||||
|
|
||||||
|
filterName = 'nlmeans_opencl' if self.__useHardware else 'nlmeans'
|
||||||
|
|
||||||
|
payload = {'identifier': NlmeansFilter.IDENTIFIER,
|
||||||
|
'parameters': {
|
||||||
|
'strength': strength,
|
||||||
|
'patch_size': patchSize,
|
||||||
|
'chroma_patch_size': chromaPatchSize,
|
||||||
|
'research_window': researchWindow,
|
||||||
|
'chroma_research_window': chromaResearchWindow
|
||||||
|
},
|
||||||
|
'suffices': suffices,
|
||||||
|
'variant': f"DS{strength}-DP{patchSize}-DPC{chromaPatchSize}"
|
||||||
|
+ f"-DR{researchWindow}-DRC{chromaResearchWindow}",
|
||||||
|
'tokens': ['-vf', f"{filterName}=s={strength}"
|
||||||
|
+ f":p={patchSize}"
|
||||||
|
+ f":pc={chromaPatchSize}"
|
||||||
|
+ f":r={researchWindow}"
|
||||||
|
+ f":rc={chromaResearchWindow}"]}
|
||||||
|
|
||||||
|
return payload
|
||||||
|
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
for it in itertools.product(self.__strengthList,
|
||||||
|
self.__patchSizeList,
|
||||||
|
self.__chromaPatchSizeList,
|
||||||
|
self.__researchWindowList,
|
||||||
|
self.__chromaResearchWindowList):
|
||||||
|
yield self.getPayload(it)
|
||||||
54
src/ffx/filter/preset_filter.py
Normal file
54
src/ffx/filter/preset_filter.py
Normal file
@@ -0,0 +1,54 @@
|
|||||||
|
import itertools
|
||||||
|
|
||||||
|
from .filter import Filter
|
||||||
|
|
||||||
|
|
||||||
|
class PresetFilter(Filter):
|
||||||
|
|
||||||
|
IDENTIFIER = 'preset'
|
||||||
|
|
||||||
|
DEFAULT_PRESET = 5
|
||||||
|
|
||||||
|
PRESET_KEY = 'preset'
|
||||||
|
|
||||||
|
def __init__(self, **kwargs):
|
||||||
|
|
||||||
|
self.__presetsList = []
|
||||||
|
presets = str(kwargs.get(PresetFilter.PRESET_KEY, ''))
|
||||||
|
if presets:
|
||||||
|
presetTokens = presets.split(',')
|
||||||
|
for q in presetTokens:
|
||||||
|
try:
|
||||||
|
presetValue = int(q)
|
||||||
|
except:
|
||||||
|
raise ValueError('PresetFilter: Preset value has to be of type int')
|
||||||
|
if presetValue < 0 or presetValue > 13:
|
||||||
|
raise ValueError('PresetFilter: Preset value has to be between 0 and 13')
|
||||||
|
self.__presetsList.append(presetValue)
|
||||||
|
else:
|
||||||
|
self.__presetsList = [PresetFilter.DEFAULT_PRESET]
|
||||||
|
|
||||||
|
super().__init__(self)
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self, preset):
|
||||||
|
|
||||||
|
suffices = []
|
||||||
|
|
||||||
|
if len(self.__presetsList) > 1:
|
||||||
|
suffices += [f"p{preset}"]
|
||||||
|
|
||||||
|
payload = {'identifier': PresetFilter.IDENTIFIER,
|
||||||
|
'parameters': {
|
||||||
|
'preset': preset
|
||||||
|
},
|
||||||
|
'suffices': suffices,
|
||||||
|
'variant': f"P{preset}",
|
||||||
|
'tokens': []}
|
||||||
|
|
||||||
|
return payload
|
||||||
|
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
for q in self.__presetsList:
|
||||||
|
yield self.getPayload(q)
|
||||||
54
src/ffx/filter/quality_filter.py
Normal file
54
src/ffx/filter/quality_filter.py
Normal file
@@ -0,0 +1,54 @@
|
|||||||
|
import itertools
|
||||||
|
|
||||||
|
from .filter import Filter
|
||||||
|
|
||||||
|
|
||||||
|
class QualityFilter(Filter):
|
||||||
|
|
||||||
|
IDENTIFIER = 'quality'
|
||||||
|
|
||||||
|
DEFAULT_QUALITY = 32
|
||||||
|
|
||||||
|
QUALITY_KEY = 'quality'
|
||||||
|
|
||||||
|
def __init__(self, **kwargs):
|
||||||
|
|
||||||
|
self.__qualitiesList = []
|
||||||
|
qualities = kwargs.get(QualityFilter.QUALITY_KEY, '')
|
||||||
|
if qualities:
|
||||||
|
qualityTokens = qualities.split(',')
|
||||||
|
for q in qualityTokens:
|
||||||
|
try:
|
||||||
|
qualityValue = int(q)
|
||||||
|
except:
|
||||||
|
raise ValueError('QualityFilter: Quality value has to be of type int')
|
||||||
|
if qualityValue < 0 or qualityValue > 63:
|
||||||
|
raise ValueError('QualityFilter: Quality value has to be between 0 and 63')
|
||||||
|
self.__qualitiesList.append(qualityValue)
|
||||||
|
else:
|
||||||
|
self.__qualitiesList = [QualityFilter.DEFAULT_QUALITY]
|
||||||
|
|
||||||
|
super().__init__(self)
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self, quality):
|
||||||
|
|
||||||
|
suffices = []
|
||||||
|
|
||||||
|
if len(self.__qualitiesList) > 1:
|
||||||
|
suffices += [f"q{quality}"]
|
||||||
|
|
||||||
|
payload = {'identifier': QualityFilter.IDENTIFIER,
|
||||||
|
'parameters': {
|
||||||
|
'quality': quality
|
||||||
|
},
|
||||||
|
'suffices': suffices,
|
||||||
|
'variant': f"Q{quality}",
|
||||||
|
'tokens': []}
|
||||||
|
|
||||||
|
return payload
|
||||||
|
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
for q in self.__qualitiesList:
|
||||||
|
yield self.getPayload(q)
|
||||||
6
src/ffx/filter/scale_filter.py
Normal file
6
src/ffx/filter/scale_filter.py
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
from .filter import Filter
|
||||||
|
|
||||||
|
class ScaleFilter(Filter):
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
super().__init__(self)
|
||||||
13
src/ffx/help_screen.py
Normal file
13
src/ffx/help_screen.py
Normal file
@@ -0,0 +1,13 @@
|
|||||||
|
from textual.app import ComposeResult
|
||||||
|
from textual.screen import Screen
|
||||||
|
from textual.widgets import Footer, Placeholder
|
||||||
|
|
||||||
|
class HelpScreen(Screen):
|
||||||
|
def __init__(self):
|
||||||
|
super().__init__()
|
||||||
|
context = self.app.getContext()
|
||||||
|
|
||||||
|
def compose(self) -> ComposeResult:
|
||||||
|
yield Placeholder("Help Screen")
|
||||||
|
yield Footer()
|
||||||
|
|
||||||
187
src/ffx/helper.py
Normal file
187
src/ffx/helper.py
Normal file
@@ -0,0 +1,187 @@
|
|||||||
|
import re, logging
|
||||||
|
|
||||||
|
from jinja2 import Environment, Undefined
|
||||||
|
from .constants import DEFAULT_OUTPUT_FILENAME_TEMPLATE
|
||||||
|
from .configuration_controller import ConfigurationController
|
||||||
|
|
||||||
|
|
||||||
|
class EmptyStringUndefined(Undefined):
|
||||||
|
def __str__(self):
|
||||||
|
return ''
|
||||||
|
|
||||||
|
|
||||||
|
DIFF_ADDED_KEY = 'added'
|
||||||
|
DIFF_REMOVED_KEY = 'removed'
|
||||||
|
DIFF_CHANGED_KEY = 'changed'
|
||||||
|
DIFF_UNCHANGED_KEY = 'unchanged'
|
||||||
|
|
||||||
|
|
||||||
|
def dictDiff(a : dict, b : dict):
|
||||||
|
|
||||||
|
a_keys = set(a.keys())
|
||||||
|
b_keys = set(b.keys())
|
||||||
|
|
||||||
|
a_only = a_keys - b_keys
|
||||||
|
b_only = b_keys - a_keys
|
||||||
|
a_b = a_keys & b_keys
|
||||||
|
|
||||||
|
changed = {k for k in a_b if a[k] != b[k]}
|
||||||
|
|
||||||
|
diffResult = {}
|
||||||
|
|
||||||
|
|
||||||
|
if a_only:
|
||||||
|
diffResult[DIFF_REMOVED_KEY] = a_only
|
||||||
|
diffResult[DIFF_UNCHANGED_KEY] = b_keys
|
||||||
|
if b_only:
|
||||||
|
diffResult[DIFF_ADDED_KEY] = b_only
|
||||||
|
if changed:
|
||||||
|
diffResult[DIFF_CHANGED_KEY] = changed
|
||||||
|
|
||||||
|
return diffResult
|
||||||
|
|
||||||
|
def dictCache(element: dict, cache: list = []):
|
||||||
|
for index in range(len(cache)):
|
||||||
|
diff = dictDiff(cache[index], element)
|
||||||
|
if not diff:
|
||||||
|
return index, cache
|
||||||
|
cache.append(element)
|
||||||
|
return -1, cache
|
||||||
|
|
||||||
|
|
||||||
|
def setDiff(a : set, b : set) -> set:
|
||||||
|
|
||||||
|
a_only = a - b
|
||||||
|
b_only = b - a
|
||||||
|
|
||||||
|
diffResult = {}
|
||||||
|
|
||||||
|
if a_only:
|
||||||
|
diffResult[DIFF_REMOVED_KEY] = a_only
|
||||||
|
if b_only:
|
||||||
|
diffResult[DIFF_ADDED_KEY] = b_only
|
||||||
|
|
||||||
|
return diffResult
|
||||||
|
|
||||||
|
|
||||||
|
def permutateList(inputList: list, permutation: list):
|
||||||
|
|
||||||
|
# 0,1,2: ABC
|
||||||
|
# 0,2,1: ACB
|
||||||
|
# 1,2,0: BCA
|
||||||
|
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def filterFilename(fileName: str) -> str:
|
||||||
|
"""This filter replaces charactes from TMDB responses with characters
|
||||||
|
less problemating when using in filenames or removes them"""
|
||||||
|
|
||||||
|
fileName = str(fileName).replace('/', '-')
|
||||||
|
fileName = str(fileName).replace(':', ';')
|
||||||
|
fileName = str(fileName).replace('*', '')
|
||||||
|
fileName = str(fileName).replace("'", '')
|
||||||
|
fileName = str(fileName).replace("?", '#')
|
||||||
|
|
||||||
|
return fileName.strip()
|
||||||
|
|
||||||
|
def substituteTmdbFilename(fileName: str) -> str:
|
||||||
|
"""If chaining this method with filterFilename use this one first as the latter will destroy some patterns"""
|
||||||
|
|
||||||
|
# This indicates filler episodes in TMDB episode names
|
||||||
|
fileName = str(fileName).replace(' (*)', '')
|
||||||
|
fileName = str(fileName).replace('(*)', '')
|
||||||
|
|
||||||
|
# This indicates the index of multi-episode files
|
||||||
|
episodePartMatch = re.search("\\(([0-9]+)\\)$", fileName)
|
||||||
|
if episodePartMatch is not None:
|
||||||
|
partSuffix = str(episodePartMatch.group(0))
|
||||||
|
partIndex = episodePartMatch.groups()[0]
|
||||||
|
fileName = str(fileName).replace(partSuffix, f"Teil {partIndex}")
|
||||||
|
|
||||||
|
# Also multi-episodes with first and last episode index
|
||||||
|
episodePartMatch = re.search("\\(([0-9]+)[-\\/]([0-9]+)\\)$", fileName)
|
||||||
|
if episodePartMatch is not None:
|
||||||
|
partSuffix = str(episodePartMatch.group(0))
|
||||||
|
partFirstIndex = episodePartMatch.groups()[0]
|
||||||
|
partLastIndex = episodePartMatch.groups()[1]
|
||||||
|
fileName = str(fileName).replace(partSuffix, f"Teil {partFirstIndex}-{partLastIndex}")
|
||||||
|
|
||||||
|
return fileName
|
||||||
|
|
||||||
|
|
||||||
|
def getEpisodeFileBasename(showName,
|
||||||
|
episodeName,
|
||||||
|
season,
|
||||||
|
episode,
|
||||||
|
indexSeasonDigits = 2,
|
||||||
|
indexEpisodeDigits = 2,
|
||||||
|
indicatorSeasonDigits = 2,
|
||||||
|
indicatorEpisodeDigits = 2,
|
||||||
|
context = None):
|
||||||
|
"""
|
||||||
|
One Piece:
|
||||||
|
indexSeasonDigits = 0,
|
||||||
|
indexEpisodeDigits = 4,
|
||||||
|
indicatorSeasonDigits = 2,
|
||||||
|
indicatorEpisodeDigits = 4
|
||||||
|
|
||||||
|
Three-Body:
|
||||||
|
indexSeasonDigits = 0,
|
||||||
|
indexEpisodeDigits = 2,
|
||||||
|
indicatorSeasonDigits = 2,
|
||||||
|
indicatorEpisodeDigits = 2
|
||||||
|
|
||||||
|
Dragonball:
|
||||||
|
indexSeasonDigits = 0,
|
||||||
|
indexEpisodeDigits = 3,
|
||||||
|
indicatorSeasonDigits = 2,
|
||||||
|
indicatorEpisodeDigits = 3
|
||||||
|
|
||||||
|
Boruto:
|
||||||
|
indexSeasonDigits = 0,
|
||||||
|
indexEpisodeDigits = 4,
|
||||||
|
indicatorSeasonDigits = 2,
|
||||||
|
indicatorEpisodeDigits = 4
|
||||||
|
"""
|
||||||
|
|
||||||
|
cc: ConfigurationController = context['config'] if context is not None and 'config' in context.keys() else None
|
||||||
|
configData = cc.getData() if cc is not None else {}
|
||||||
|
outputFilenameTemplate = configData.get(ConfigurationController.OUTPUT_FILENAME_TEMPLATE_KEY,
|
||||||
|
DEFAULT_OUTPUT_FILENAME_TEMPLATE)
|
||||||
|
|
||||||
|
if context is not None and 'logger' in context.keys():
|
||||||
|
logger = context['logger']
|
||||||
|
else:
|
||||||
|
logger = logging.getLogger('FFX')
|
||||||
|
logger.addHandler(logging.NullHandler())
|
||||||
|
|
||||||
|
|
||||||
|
indexSeparator = ' ' if indexSeasonDigits or indexEpisodeDigits else ''
|
||||||
|
seasonIndex = '{num:{fill}{width}}'.format(num=season, fill='0', width=indexSeasonDigits) if indexSeasonDigits else ''
|
||||||
|
episodeIndex = '{num:{fill}{width}}'.format(num=episode, fill='0', width=indexEpisodeDigits) if indexEpisodeDigits else ''
|
||||||
|
|
||||||
|
indicatorSeparator = ' - ' if indicatorSeasonDigits or indicatorEpisodeDigits else ''
|
||||||
|
seasonIndicator = 'S{num:{fill}{width}}'.format(num=season, fill='0', width=indicatorSeasonDigits) if indicatorSeasonDigits else ''
|
||||||
|
episodeIndicator = 'E{num:{fill}{width}}'.format(num=episode, fill='0', width=indicatorEpisodeDigits) if indicatorEpisodeDigits else ''
|
||||||
|
|
||||||
|
jinjaKwargs = {
|
||||||
|
'ffx_show_name': showName,
|
||||||
|
'ffx_index_separator': indexSeparator,
|
||||||
|
'ffx_season_index': str(seasonIndex),
|
||||||
|
'ffx_episode_index': str(episodeIndex),
|
||||||
|
'ffx_index': str(seasonIndex) + str(episodeIndex),
|
||||||
|
'ffx_episode_name': episodeName,
|
||||||
|
'ffx_indicator_separator': indicatorSeparator,
|
||||||
|
'ffx_season_indicator': str(seasonIndicator),
|
||||||
|
'ffx_episode_indicator': str(episodeIndicator),
|
||||||
|
'ffx_indicator': str(seasonIndicator) + str(episodeIndicator)
|
||||||
|
}
|
||||||
|
|
||||||
|
jinjaEnv = Environment(undefined=EmptyStringUndefined)
|
||||||
|
jinjaTemplate = jinjaEnv.from_string(outputFilenameTemplate)
|
||||||
|
return jinjaTemplate.render(**jinjaKwargs)
|
||||||
|
|
||||||
|
# return ''.join(filenameTokens)
|
||||||
|
|
||||||
106
src/ffx/iso_language.py
Normal file
106
src/ffx/iso_language.py
Normal file
@@ -0,0 +1,106 @@
|
|||||||
|
from enum import Enum
|
||||||
|
import difflib
|
||||||
|
|
||||||
|
class IsoLanguage(Enum):
|
||||||
|
|
||||||
|
AFRIKAANS = {"name": "Afrikaans", "iso639_1": "af", "iso639_2": "afr"}
|
||||||
|
ALBANIAN = {"name": "Albanian", "iso639_1": "sq", "iso639_2": "alb"}
|
||||||
|
ARABIC = {"name": "Arabic", "iso639_1": "ar", "iso639_2": "ara"}
|
||||||
|
ARMENIAN = {"name": "Armenian", "iso639_1": "hy", "iso639_2": "arm"}
|
||||||
|
AZERBAIJANI = {"name": "Azerbaijani", "iso639_1": "az", "iso639_2": "aze"}
|
||||||
|
BASQUE = {"name": "Basque", "iso639_1": "eu", "iso639_2": "baq"}
|
||||||
|
BELARUSIAN = {"name": "Belarusian", "iso639_1": "be", "iso639_2": "bel"}
|
||||||
|
BULGARIAN = {"name": "Bulgarian", "iso639_1": "bg", "iso639_2": "bul"}
|
||||||
|
CATALAN = {"name": "Catalan", "iso639_1": "ca", "iso639_2": "cat"}
|
||||||
|
CHINESE = {"name": "Chinese", "iso639_1": "zh", "iso639_2": "chi"}
|
||||||
|
CROATIAN = {"name": "Croatian", "iso639_1": "hr", "iso639_2": "hrv"}
|
||||||
|
CZECH = {"name": "Czech", "iso639_1": "cs", "iso639_2": "cze"}
|
||||||
|
DANISH = {"name": "Danish", "iso639_1": "da", "iso639_2": "dan"}
|
||||||
|
DUTCH = {"name": "Dutch", "iso639_1": "nl", "iso639_2": "dut"}
|
||||||
|
ENGLISH = {"name": "English", "iso639_1": "en", "iso639_2": "eng"}
|
||||||
|
ESTONIAN = {"name": "Estonian", "iso639_1": "et", "iso639_2": "est"}
|
||||||
|
FINNISH = {"name": "Finnish", "iso639_1": "fi", "iso639_2": "fin"}
|
||||||
|
FRENCH = {"name": "French", "iso639_1": "fr", "iso639_2": "fre"}
|
||||||
|
GEORGIAN = {"name": "Georgian", "iso639_1": "ka", "iso639_2": "geo"}
|
||||||
|
GERMAN = {"name": "German", "iso639_1": "de", "iso639_2": "ger"}
|
||||||
|
GREEK = {"name": "Greek", "iso639_1": "el", "iso639_2": "gre"}
|
||||||
|
HEBREW = {"name": "Hebrew", "iso639_1": "he", "iso639_2": "heb"}
|
||||||
|
HINDI = {"name": "Hindi", "iso639_1": "hi", "iso639_2": "hin"}
|
||||||
|
HUNGARIAN = {"name": "Hungarian", "iso639_1": "hu", "iso639_2": "hun"}
|
||||||
|
ICELANDIC = {"name": "Icelandic", "iso639_1": "is", "iso639_2": "ice"}
|
||||||
|
INDONESIAN = {"name": "Indonesian", "iso639_1": "id", "iso639_2": "ind"}
|
||||||
|
IRISH = {"name": "Irish", "iso639_1": "ga", "iso639_2": "gle"}
|
||||||
|
ITALIAN = {"name": "Italian", "iso639_1": "it", "iso639_2": "ita"}
|
||||||
|
JAPANESE = {"name": "Japanese", "iso639_1": "ja", "iso639_2": "jpn"}
|
||||||
|
KAZAKH = {"name": "Kazakh", "iso639_1": "kk", "iso639_2": "kaz"}
|
||||||
|
KOREAN = {"name": "Korean", "iso639_1": "ko", "iso639_2": "kor"}
|
||||||
|
LATIN = {"name": "Latin", "iso639_1": "la", "iso639_2": "lat"}
|
||||||
|
LATVIAN = {"name": "Latvian", "iso639_1": "lv", "iso639_2": "lav"}
|
||||||
|
LITHUANIAN = {"name": "Lithuanian", "iso639_1": "lt", "iso639_2": "lit"}
|
||||||
|
MACEDONIAN = {"name": "Macedonian", "iso639_1": "mk", "iso639_2": "mac"}
|
||||||
|
MALAY = {"name": "Malay", "iso639_1": "ms", "iso639_2": "may"}
|
||||||
|
MALTESE = {"name": "Maltese", "iso639_1": "mt", "iso639_2": "mlt"}
|
||||||
|
NORWEGIAN = {"name": "Norwegian", "iso639_1": "no", "iso639_2": "nor"}
|
||||||
|
PERSIAN = {"name": "Persian", "iso639_1": "fa", "iso639_2": "per"}
|
||||||
|
POLISH = {"name": "Polish", "iso639_1": "pl", "iso639_2": "pol"}
|
||||||
|
PORTUGUESE = {"name": "Portuguese", "iso639_1": "pt", "iso639_2": "por"}
|
||||||
|
ROMANIAN = {"name": "Romanian", "iso639_1": "ro", "iso639_2": "rum"}
|
||||||
|
RUSSIAN = {"name": "Russian", "iso639_1": "ru", "iso639_2": "rus"}
|
||||||
|
NORTHERN_SAMI = {"name": "Northern Sami", "iso639_1": "se", "iso639_2": "sme"}
|
||||||
|
SAMOAN = {"name": "Samoan", "iso639_1": "sm", "iso639_2": "smo"}
|
||||||
|
SANGO = {"name": "Sango", "iso639_1": "sg", "iso639_2": "sag"}
|
||||||
|
SANSKRIT = {"name": "Sanskrit", "iso639_1": "sa", "iso639_2": "san"}
|
||||||
|
SARDINIAN = {"name": "Sardinian", "iso639_1": "sc", "iso639_2": "srd"}
|
||||||
|
SERBIAN = {"name": "Serbian", "iso639_1": "sr", "iso639_2": "srp"}
|
||||||
|
SHONA = {"name": "Shona", "iso639_1": "sn", "iso639_2": "sna"}
|
||||||
|
SINDHI = {"name": "Sindhi", "iso639_1": "sd", "iso639_2": "snd"}
|
||||||
|
SINHALA = {"name": "Sinhala", "iso639_1": "si", "iso639_2": "sin"}
|
||||||
|
SLOVAK = {"name": "Slovak", "iso639_1": "sk", "iso639_2": "slk"}
|
||||||
|
SLOVENIAN = {"name": "Slovenian", "iso639_1": "sl", "iso639_2": "slv"}
|
||||||
|
SOMALI = {"name": "Somali", "iso639_1": "so", "iso639_2": "som"}
|
||||||
|
SOUTHERN_SOTHO = {"name": "Southern Sotho", "iso639_1": "st", "iso639_2": "sot"}
|
||||||
|
SPANISH = {"name": "Spanish", "iso639_1": "es", "iso639_2": "spa"}
|
||||||
|
SUNDANESE = {"name": "Sundanese", "iso639_1": "su", "iso639_2": "sun"}
|
||||||
|
SWAHILI = {"name": "Swahili", "iso639_1": "sw", "iso639_2": "swa"}
|
||||||
|
SWATI = {"name": "Swati", "iso639_1": "ss", "iso639_2": "ssw"}
|
||||||
|
SWEDISH = {"name": "Swedish", "iso639_1": "sv", "iso639_2": "swe"}
|
||||||
|
TAGALOG = {"name": "Tagalog", "iso639_1": "tl", "iso639_2": "tgl"}
|
||||||
|
TAMIL = {"name": "Tamil", "iso639_1": "ta", "iso639_2": "tam"}
|
||||||
|
THAI = {"name": "Thai", "iso639_1": "th", "iso639_2": "tha"}
|
||||||
|
TURKISH = {"name": "Turkish", "iso639_1": "tr", "iso639_2": "tur"}
|
||||||
|
UKRAINIAN = {"name": "Ukrainian", "iso639_1": "uk", "iso639_2": "ukr"}
|
||||||
|
URDU = {"name": "Urdu", "iso639_1": "ur", "iso639_2": "urd"}
|
||||||
|
VIETNAMESE = {"name": "Vietnamese", "iso639_1": "vi", "iso639_2": "vie"}
|
||||||
|
WELSH = {"name": "Welsh", "iso639_1": "cy", "iso639_2": "wel"}
|
||||||
|
|
||||||
|
UNDEFINED = {"name": "undefined", "iso639_1": "xx", "iso639_2": "und"}
|
||||||
|
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def find(label : str):
|
||||||
|
|
||||||
|
closestMatches = difflib.get_close_matches(label, [l.value["name"] for l in IsoLanguage], n=1)
|
||||||
|
|
||||||
|
if closestMatches:
|
||||||
|
foundLangs = [l for l in IsoLanguage if l.value['name'] == closestMatches[0]]
|
||||||
|
return foundLangs[0] if foundLangs else IsoLanguage.UNDEFINED
|
||||||
|
else:
|
||||||
|
return IsoLanguage.UNDEFINED
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def findThreeLetter(theeLetter : str):
|
||||||
|
foundLangs = [l for l in IsoLanguage if l.value['iso639_2'] == str(theeLetter)]
|
||||||
|
return foundLangs[0] if foundLangs else IsoLanguage.UNDEFINED
|
||||||
|
|
||||||
|
|
||||||
|
def label(self):
|
||||||
|
return str(self.value['name'])
|
||||||
|
|
||||||
|
def twoLetter(self):
|
||||||
|
return str(self.value['iso639_1'])
|
||||||
|
|
||||||
|
def threeLetter(self):
|
||||||
|
return str(self.value['iso639_2'])
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
47
src/ffx/media_controller.py
Normal file
47
src/ffx/media_controller.py
Normal file
@@ -0,0 +1,47 @@
|
|||||||
|
import click
|
||||||
|
|
||||||
|
from ffx.model.pattern import Pattern
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from ffx.tag_controller import TagController
|
||||||
|
from ffx.track_controller import TrackController
|
||||||
|
|
||||||
|
class MediaController():
|
||||||
|
|
||||||
|
def __init__(self, context):
|
||||||
|
|
||||||
|
self.context = context
|
||||||
|
self.Session = self.context['database']['session'] # convenience
|
||||||
|
|
||||||
|
self.__logger = context['logger']
|
||||||
|
|
||||||
|
self.__tc = TrackController(context = context)
|
||||||
|
self.__tac = TagController(context = context)
|
||||||
|
|
||||||
|
def setPatternMediaDescriptor(self, mediaDescriptor: MediaDescriptor, patternId: int):
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
pid = int(patternId)
|
||||||
|
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Pattern).filter(Pattern.id == pid)
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
pattern = q.first
|
||||||
|
|
||||||
|
for mediaTagKey, mediaTagValue in mediaDescriptor.getTags():
|
||||||
|
self.__tac.updateMediaTag(pid, mediaTagKey, mediaTagValue)
|
||||||
|
for trackDescriptor in mediaDescriptor.getAllTrackDescriptors():
|
||||||
|
self.__tc.addTrack(trackDescriptor, patternId = pid)
|
||||||
|
|
||||||
|
s.commit()
|
||||||
|
return True
|
||||||
|
else:
|
||||||
|
return False
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
self.__logger.error(f"MediaController.setPatternMediaDescriptor(): {repr(ex)}")
|
||||||
|
raise click.ClickException(f"MediaController.setPatternMediaDescriptor(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
547
src/ffx/media_descriptor.py
Normal file
547
src/ffx/media_descriptor.py
Normal file
@@ -0,0 +1,547 @@
|
|||||||
|
import os, re, click, logging
|
||||||
|
|
||||||
|
from typing import List, Self
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
from ffx.iso_language import IsoLanguage
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from ffx.track_codec import TrackCodec
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
|
||||||
|
from ffx.helper import dictDiff, DIFF_ADDED_KEY, DIFF_CHANGED_KEY, DIFF_REMOVED_KEY
|
||||||
|
|
||||||
|
|
||||||
|
class MediaDescriptor:
|
||||||
|
"""This class represents the structural content of a media file including streams and metadata"""
|
||||||
|
|
||||||
|
CONTEXT_KEY = "context"
|
||||||
|
|
||||||
|
TAGS_KEY = "tags"
|
||||||
|
TRACKS_KEY = "tracks"
|
||||||
|
|
||||||
|
TRACK_DESCRIPTOR_LIST_KEY = "track_descriptors"
|
||||||
|
CLEAR_TAGS_FLAG_KEY = "clear_tags"
|
||||||
|
|
||||||
|
FFPROBE_DISPOSITION_KEY = "disposition"
|
||||||
|
FFPROBE_TAGS_KEY = "tags"
|
||||||
|
FFPROBE_CODEC_TYPE_KEY = "codec_type"
|
||||||
|
|
||||||
|
#407 remove as well
|
||||||
|
EXCLUDED_MEDIA_TAGS = ["creation_time"]
|
||||||
|
|
||||||
|
SEASON_EPISODE_STREAM_LANGUAGE_MATCH = '[sS]([0-9]+)[eE]([0-9]+)_([0-9]+)_([a-z]{3})(?:_([A-Z]{3}))*'
|
||||||
|
SUBTITLE_FILE_EXTENSION = 'vtt'
|
||||||
|
|
||||||
|
def __init__(self, **kwargs):
|
||||||
|
|
||||||
|
if MediaDescriptor.CONTEXT_KEY in kwargs.keys():
|
||||||
|
if type(kwargs[MediaDescriptor.CONTEXT_KEY]) is not dict:
|
||||||
|
raise TypeError(
|
||||||
|
f"MediaDescriptor.__init__(): Argument {MediaDescriptor.CONTEXT_KEY} is required to be of type dict"
|
||||||
|
)
|
||||||
|
self.__context = kwargs[MediaDescriptor.CONTEXT_KEY]
|
||||||
|
self.__logger = self.__context['logger']
|
||||||
|
else:
|
||||||
|
self.__context = {}
|
||||||
|
self.__logger = logging.getLogger('FFX')
|
||||||
|
self.__logger.addHandler(logging.NullHandler())
|
||||||
|
|
||||||
|
if MediaDescriptor.TAGS_KEY in kwargs.keys():
|
||||||
|
if type(kwargs[MediaDescriptor.TAGS_KEY]) is not dict:
|
||||||
|
raise TypeError(
|
||||||
|
f"MediaDescriptor.__init__(): Argument {MediaDescriptor.TAGS_KEY} is required to be of type dict"
|
||||||
|
)
|
||||||
|
self.__mediaTags = kwargs[MediaDescriptor.TAGS_KEY]
|
||||||
|
else:
|
||||||
|
self.__mediaTags = {}
|
||||||
|
|
||||||
|
if MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY in kwargs.keys():
|
||||||
|
if (
|
||||||
|
type(kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY]) is not list
|
||||||
|
): # Use List typehint for TrackDescriptor as well if it works
|
||||||
|
raise TypeError(
|
||||||
|
f"MediaDescriptor.__init__(): Argument {MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY} is required to be of type list"
|
||||||
|
)
|
||||||
|
for d in kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY]:
|
||||||
|
if type(d) is not TrackDescriptor:
|
||||||
|
raise TypeError(
|
||||||
|
f"TrackDesciptor.__init__(): All elements of argument list {MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY} are required to be of type TrackDescriptor"
|
||||||
|
)
|
||||||
|
self.__trackDescriptors = kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY]
|
||||||
|
else:
|
||||||
|
self.__trackDescriptors = []
|
||||||
|
|
||||||
|
def setTrackLanguage(self, language: str, index: int, trackType: TrackType = None):
|
||||||
|
|
||||||
|
trackLanguage = IsoLanguage.findThreeLetter(language)
|
||||||
|
if trackLanguage == IsoLanguage.UNDEFINED:
|
||||||
|
self.__logger.warning('MediaDescriptor.setTrackLanguage(): Parameter language does not contain a registered '
|
||||||
|
+ f"ISO 639 3-letter language code, skipping to set language for"
|
||||||
|
+ str('' if trackType is None else trackType.label()) + f"track {index}")
|
||||||
|
|
||||||
|
trackList = self.getTrackDescriptors(trackType=trackType)
|
||||||
|
|
||||||
|
if index < 0 or index > len(trackList) - 1:
|
||||||
|
self.__logger.warning(f"MediaDescriptor.setTrackLanguage(): Parameter index ({index}) is "
|
||||||
|
+ f"out of range of {'' if trackType is None else trackType.label()}track list")
|
||||||
|
|
||||||
|
td: TrackDescriptor = trackList[index]
|
||||||
|
td.setLanguage(trackLanguage)
|
||||||
|
|
||||||
|
return
|
||||||
|
|
||||||
|
|
||||||
|
def setTrackTitle(self, title: str, index: int, trackType: TrackType = None):
|
||||||
|
|
||||||
|
trackList = self.getTrackDescriptors(trackType=trackType)
|
||||||
|
|
||||||
|
if index < 0 or index > len(trackList) - 1:
|
||||||
|
self.__logger.error(f"MediaDescriptor.setTrackTitle(): Parameter index ({index}) is "
|
||||||
|
+ f"out of range of {'' if trackType is None else trackType.label()}track list")
|
||||||
|
raise click.Abort()
|
||||||
|
|
||||||
|
td: TrackDescriptor = trackList[index]
|
||||||
|
td.setTitle(title)
|
||||||
|
|
||||||
|
|
||||||
|
def setDefaultSubTrack(self, trackType: TrackType, subIndex: int):
|
||||||
|
for t in self.getAllTrackDescriptors():
|
||||||
|
if t.getType() == trackType:
|
||||||
|
t.setDispositionFlag(
|
||||||
|
TrackDisposition.DEFAULT, t.getSubIndex() == int(subIndex)
|
||||||
|
)
|
||||||
|
|
||||||
|
def setForcedSubTrack(self, trackType: TrackType, subIndex: int):
|
||||||
|
for t in self.getAllTrackDescriptors():
|
||||||
|
if t.getType() == trackType:
|
||||||
|
t.setDispositionFlag(
|
||||||
|
TrackDisposition.FORCED, t.getSubIndex() == int(subIndex)
|
||||||
|
)
|
||||||
|
|
||||||
|
def checkConfiguration(self):
|
||||||
|
|
||||||
|
videoTracks = self.getVideoTracks()
|
||||||
|
audioTracks = self.getAudioTracks()
|
||||||
|
subtitleTracks = self.getSubtitleTracks()
|
||||||
|
|
||||||
|
if len([v for v in videoTracks if v.getDispositionFlag(TrackDisposition.DEFAULT)]) > 1:
|
||||||
|
raise ValueError('More than one default video track')
|
||||||
|
if len([a for a in audioTracks if a.getDispositionFlag(TrackDisposition.DEFAULT)]) > 1:
|
||||||
|
raise ValueError('More than one default audio track')
|
||||||
|
if len([s for s in subtitleTracks if s.getDispositionFlag(TrackDisposition.DEFAULT)]) > 1:
|
||||||
|
raise ValueError('More than one default subtitle track')
|
||||||
|
|
||||||
|
if len([v for v in videoTracks if v.getDispositionFlag(TrackDisposition.FORCED)]) > 1:
|
||||||
|
raise ValueError('More than one forced video track')
|
||||||
|
if len([a for a in audioTracks if a.getDispositionFlag(TrackDisposition.FORCED)]) > 1:
|
||||||
|
raise ValueError('More than one forced audio track')
|
||||||
|
if len([s for s in subtitleTracks if s.getDispositionFlag(TrackDisposition.FORCED)]) > 1:
|
||||||
|
raise ValueError('More than one forced subtitle track')
|
||||||
|
|
||||||
|
trackDescriptors = videoTracks + audioTracks + subtitleTracks
|
||||||
|
sourceIndices = [
|
||||||
|
t.getSourceIndex() for t in trackDescriptors
|
||||||
|
]
|
||||||
|
if len(set(sourceIndices)) < len(trackDescriptors):
|
||||||
|
raise ValueError('Multiple streams originating from the same source stream')
|
||||||
|
|
||||||
|
|
||||||
|
def applyOverrides(self, overrides: dict):
|
||||||
|
|
||||||
|
if 'languages' in overrides.keys():
|
||||||
|
for trackIndex in overrides['languages'].keys():
|
||||||
|
self.setTrackLanguage(overrides['languages'][trackIndex], trackIndex)
|
||||||
|
|
||||||
|
if 'titles' in overrides.keys():
|
||||||
|
for trackIndex in overrides['titles'].keys():
|
||||||
|
self.setTrackTitle(overrides['titles'][trackIndex], trackIndex)
|
||||||
|
|
||||||
|
if 'forced_video' in overrides.keys():
|
||||||
|
sti = int(overrides['forced_video'])
|
||||||
|
self.setForcedSubTrack(TrackType.VIDEO, sti)
|
||||||
|
self.setDefaultSubTrack(TrackType.VIDEO, sti)
|
||||||
|
|
||||||
|
elif 'default_video' in overrides.keys():
|
||||||
|
sti = int(overrides['default_video'])
|
||||||
|
self.setDefaultSubTrack(TrackType.VIDEO, sti)
|
||||||
|
|
||||||
|
if 'forced_audio' in overrides.keys():
|
||||||
|
sti = int(overrides['forced_audio'])
|
||||||
|
self.setForcedSubTrack(TrackType.AUDIO, sti)
|
||||||
|
self.setDefaultSubTrack(TrackType.AUDIO, sti)
|
||||||
|
|
||||||
|
elif 'default_audio' in overrides.keys():
|
||||||
|
sti = int(overrides['default_audio'])
|
||||||
|
self.setDefaultSubTrack(TrackType.AUDIO, sti)
|
||||||
|
|
||||||
|
if 'forced_subtitle' in overrides.keys():
|
||||||
|
sti = int(overrides['forced_subtitle'])
|
||||||
|
self.setForcedSubTrack(TrackType.SUBTITLE, sti)
|
||||||
|
self.setDefaultSubTrack(TrackType.SUBTITLE, sti)
|
||||||
|
|
||||||
|
elif 'default_subtitle' in overrides.keys():
|
||||||
|
sti = int(overrides['default_subtitle'])
|
||||||
|
self.setDefaultSubTrack(TrackType.SUBTITLE, sti)
|
||||||
|
|
||||||
|
if 'stream_order' in overrides.keys():
|
||||||
|
self.rearrangeTrackDescriptors(overrides['stream_order'])
|
||||||
|
|
||||||
|
|
||||||
|
def applySourceIndices(self, sourceMediaDescriptor: Self):
|
||||||
|
sourceTrackDescriptors = sourceMediaDescriptor.getAllTrackDescriptors()
|
||||||
|
|
||||||
|
numTrackDescriptors = len(self.__trackDescriptors)
|
||||||
|
if len(sourceTrackDescriptors) != numTrackDescriptors:
|
||||||
|
raise ValueError('MediaDescriptor.applySourceIndices (): Number of track descriptors does not match')
|
||||||
|
|
||||||
|
for trackIndex in range(numTrackDescriptors):
|
||||||
|
self.__trackDescriptors[trackIndex].setSourceIndex(sourceTrackDescriptors[trackIndex].getSourceIndex())
|
||||||
|
|
||||||
|
|
||||||
|
def rearrangeTrackDescriptors(self, newOrder: List[int]):
|
||||||
|
if len(newOrder) != len(self.__trackDescriptors):
|
||||||
|
raise ValueError('Length of list with reordered indices does not match number of track descriptors')
|
||||||
|
reorderedTrackDescriptors = {}
|
||||||
|
for oldIndex in newOrder:
|
||||||
|
reorderedTrackDescriptors.append(self.__trackDescriptors[oldIndex])
|
||||||
|
self.__trackDescriptors = reorderedTrackDescriptors
|
||||||
|
self.reindexSubIndices()
|
||||||
|
self.reindexIndices()
|
||||||
|
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def fromFfprobe(cls, context, formatData, streamData):
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
|
||||||
|
kwargs[MediaDescriptor.CONTEXT_KEY] = context
|
||||||
|
|
||||||
|
if MediaDescriptor.FFPROBE_TAGS_KEY in formatData.keys():
|
||||||
|
kwargs[MediaDescriptor.TAGS_KEY] = formatData[
|
||||||
|
MediaDescriptor.FFPROBE_TAGS_KEY
|
||||||
|
]
|
||||||
|
|
||||||
|
kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY] = []
|
||||||
|
|
||||||
|
# TODO: Evtl obsolet
|
||||||
|
subIndexCounters = {}
|
||||||
|
|
||||||
|
for streamObj in streamData:
|
||||||
|
|
||||||
|
ffprobeCodecType = streamObj[MediaDescriptor.FFPROBE_CODEC_TYPE_KEY]
|
||||||
|
trackType = TrackType.fromLabel(ffprobeCodecType)
|
||||||
|
|
||||||
|
if trackType != TrackType.UNKNOWN:
|
||||||
|
|
||||||
|
if trackType not in subIndexCounters.keys():
|
||||||
|
subIndexCounters[trackType] = 0
|
||||||
|
|
||||||
|
kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY].append(
|
||||||
|
TrackDescriptor.fromFfprobe(
|
||||||
|
streamObj, subIndex=subIndexCounters[trackType]
|
||||||
|
)
|
||||||
|
)
|
||||||
|
subIndexCounters[trackType] += 1
|
||||||
|
|
||||||
|
return cls(**kwargs)
|
||||||
|
|
||||||
|
def getTags(self):
|
||||||
|
return self.__mediaTags
|
||||||
|
|
||||||
|
|
||||||
|
def sortSubIndices(
|
||||||
|
self, descriptors: List[TrackDescriptor]
|
||||||
|
) -> List[TrackDescriptor]:
|
||||||
|
subIndex = 0
|
||||||
|
for d in descriptors:
|
||||||
|
d.setSubIndex(subIndex)
|
||||||
|
subIndex += 1
|
||||||
|
return descriptors
|
||||||
|
|
||||||
|
def reindexSubIndices(self, trackDescriptors: list = []):
|
||||||
|
tdList = trackDescriptors if trackDescriptors else self.__trackDescriptors
|
||||||
|
subIndexCounter = {}
|
||||||
|
for td in tdList:
|
||||||
|
trackType = td.getType()
|
||||||
|
if trackType not in subIndexCounter.keys():
|
||||||
|
subIndexCounter[trackType] = 0
|
||||||
|
td.setSubIndex(subIndexCounter[trackType])
|
||||||
|
subIndexCounter[trackType] += 1
|
||||||
|
|
||||||
|
def sortIndices(
|
||||||
|
self, descriptors: List[TrackDescriptor]
|
||||||
|
) -> List[TrackDescriptor]:
|
||||||
|
index = 0
|
||||||
|
for d in descriptors:
|
||||||
|
d.setIndex(index)
|
||||||
|
index += 1
|
||||||
|
return descriptors
|
||||||
|
|
||||||
|
def reindexIndices(self, trackDescriptors: list = []):
|
||||||
|
tdList = trackDescriptors if trackDescriptors else self.__trackDescriptors
|
||||||
|
for trackIndex in range(len(tdList)):
|
||||||
|
tdList[trackIndex].setIndex(trackIndex)
|
||||||
|
|
||||||
|
|
||||||
|
def getAllTrackDescriptors(self):
|
||||||
|
"""Returns all track descriptors sorted by type: video, audio then subtitles"""
|
||||||
|
return self.getVideoTracks() + self.getAudioTracks() + self.getSubtitleTracks()
|
||||||
|
|
||||||
|
|
||||||
|
def getTrackDescriptors(self,
|
||||||
|
trackType: TrackType = None) -> List[TrackDescriptor]:
|
||||||
|
|
||||||
|
if trackType is None:
|
||||||
|
return self.__trackDescriptors
|
||||||
|
|
||||||
|
descriptorList = []
|
||||||
|
for td in self.__trackDescriptors:
|
||||||
|
if td.getType() == trackType:
|
||||||
|
descriptorList.append(td)
|
||||||
|
|
||||||
|
return descriptorList
|
||||||
|
|
||||||
|
|
||||||
|
def getVideoTracks(self) -> List[TrackDescriptor]:
|
||||||
|
return [v for v in self.__trackDescriptors if v.getType() == TrackType.VIDEO]
|
||||||
|
|
||||||
|
def getAudioTracks(self) -> List[TrackDescriptor]:
|
||||||
|
return [a for a in self.__trackDescriptors if a.getType() == TrackType.AUDIO]
|
||||||
|
|
||||||
|
def getSubtitleTracks(self) -> List[TrackDescriptor]:
|
||||||
|
return [
|
||||||
|
s
|
||||||
|
for s in self.__trackDescriptors
|
||||||
|
if s.getType() == TrackType.SUBTITLE
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
def compare(self, vsMediaDescriptor: Self):
|
||||||
|
|
||||||
|
if not isinstance(vsMediaDescriptor, self.__class__):
|
||||||
|
self.__logger.error(f"MediaDescriptor.compare(): Argument is required to be of type {self.__class__}")
|
||||||
|
raise click.Abort()
|
||||||
|
|
||||||
|
vsTags = vsMediaDescriptor.getTags()
|
||||||
|
tags = self.getTags()
|
||||||
|
|
||||||
|
# HINT: Some tags differ per file, for example creation_time, so these are removed before diff
|
||||||
|
for emt in MediaDescriptor.EXCLUDED_MEDIA_TAGS:
|
||||||
|
if emt in tags.keys():
|
||||||
|
del tags[emt]
|
||||||
|
if emt in vsTags.keys():
|
||||||
|
del vsTags[emt]
|
||||||
|
|
||||||
|
tagsDiff = dictDiff(vsTags, tags)
|
||||||
|
|
||||||
|
compareResult = {}
|
||||||
|
|
||||||
|
if tagsDiff:
|
||||||
|
compareResult[MediaDescriptor.TAGS_KEY] = tagsDiff
|
||||||
|
|
||||||
|
# Target track configuration (from DB)
|
||||||
|
# tracks = self.getAllTrackDescriptors()
|
||||||
|
tracks = self.getAllTrackDescriptors() # filtern
|
||||||
|
numTracks = len(tracks)
|
||||||
|
|
||||||
|
# Current track configuration (of file)
|
||||||
|
vsTracks = vsMediaDescriptor.getAllTrackDescriptors()
|
||||||
|
numVsTracks = len(vsTracks)
|
||||||
|
|
||||||
|
maxNumOfTracks = max(numVsTracks, numTracks)
|
||||||
|
|
||||||
|
trackCompareResult = {}
|
||||||
|
|
||||||
|
for tp in range(maxNumOfTracks):
|
||||||
|
|
||||||
|
#!
|
||||||
|
vsTrackIndex = tracks[tp].getSourceIndex()
|
||||||
|
|
||||||
|
# Will trigger if tracks are missing in file
|
||||||
|
if tp > (numVsTracks - 1):
|
||||||
|
if DIFF_ADDED_KEY not in trackCompareResult.keys():
|
||||||
|
trackCompareResult[DIFF_ADDED_KEY] = set()
|
||||||
|
trackCompareResult[DIFF_ADDED_KEY].add(tracks[tp].getIndex())
|
||||||
|
continue
|
||||||
|
|
||||||
|
# Will trigger if tracks are missing in DB definition
|
||||||
|
# New tracks will be added per update via this way
|
||||||
|
if tp > (numTracks - 1):
|
||||||
|
if DIFF_REMOVED_KEY not in trackCompareResult.keys():
|
||||||
|
trackCompareResult[DIFF_REMOVED_KEY] = {}
|
||||||
|
trackCompareResult[DIFF_REMOVED_KEY][
|
||||||
|
vsTracks[vsTrackIndex].getIndex()
|
||||||
|
] = vsTracks[vsTrackIndex]
|
||||||
|
continue
|
||||||
|
|
||||||
|
# assumption is made here that the track order will not change for all files of a sequence
|
||||||
|
trackDiff = tracks[tp].compare(vsTracks[vsTrackIndex])
|
||||||
|
|
||||||
|
if trackDiff:
|
||||||
|
if DIFF_CHANGED_KEY not in trackCompareResult.keys():
|
||||||
|
trackCompareResult[DIFF_CHANGED_KEY] = {}
|
||||||
|
trackCompareResult[DIFF_CHANGED_KEY][
|
||||||
|
vsTracks[vsTrackIndex].getIndex()
|
||||||
|
] = trackDiff
|
||||||
|
|
||||||
|
if trackCompareResult:
|
||||||
|
compareResult[MediaDescriptor.TRACKS_KEY] = trackCompareResult
|
||||||
|
|
||||||
|
return compareResult
|
||||||
|
|
||||||
|
|
||||||
|
def getImportFileTokens(self, use_sub_index: bool = True):
|
||||||
|
|
||||||
|
importFileTokens = []
|
||||||
|
|
||||||
|
td: TrackDescriptor
|
||||||
|
for td in self.__trackDescriptors:
|
||||||
|
|
||||||
|
importedFilePath = td.getExternalSourceFilePath()
|
||||||
|
|
||||||
|
if importedFilePath:
|
||||||
|
|
||||||
|
self.__logger.info(f"Substituting subtitle stream #{td.getIndex()} "
|
||||||
|
+ f"({td.getType().label()}:{td.getSubIndex()}) "
|
||||||
|
+ f"with import from file {td.getExternalSourceFilePath()}")
|
||||||
|
|
||||||
|
importFileTokens += [
|
||||||
|
"-i",
|
||||||
|
importedFilePath,
|
||||||
|
]
|
||||||
|
|
||||||
|
return importFileTokens
|
||||||
|
|
||||||
|
|
||||||
|
def getInputMappingTokens(self, use_sub_index: bool = True, only_video: bool = False):
|
||||||
|
"""Tracks must be reordered for source index order"""
|
||||||
|
|
||||||
|
inputMappingTokens = []
|
||||||
|
|
||||||
|
filePointer = 1
|
||||||
|
for trackIndex in range(len(self.__trackDescriptors)):
|
||||||
|
|
||||||
|
td = self.__trackDescriptors[trackIndex]
|
||||||
|
|
||||||
|
stdi = self.__trackDescriptors[td.getSourceIndex()].getIndex()
|
||||||
|
stdsi = self.__trackDescriptors[td.getSourceIndex()].getSubIndex()
|
||||||
|
|
||||||
|
# sti = self.__trackDescriptors[trackIndex].getSourceIndex()
|
||||||
|
# sotd = sourceOrderTrackDescriptors[sti]
|
||||||
|
|
||||||
|
trackType = td.getType()
|
||||||
|
|
||||||
|
if (trackType == TrackType.VIDEO or not only_video):
|
||||||
|
|
||||||
|
importedFilePath = td.getExternalSourceFilePath()
|
||||||
|
|
||||||
|
if use_sub_index:
|
||||||
|
|
||||||
|
if importedFilePath:
|
||||||
|
|
||||||
|
inputMappingTokens += [
|
||||||
|
"-map",
|
||||||
|
f"{filePointer}:{trackType.indicator()}:0",
|
||||||
|
]
|
||||||
|
filePointer += 1
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
if td.getCodec() != TrackCodec.PGS:
|
||||||
|
inputMappingTokens += [
|
||||||
|
"-map",
|
||||||
|
f"0:{trackType.indicator()}:{stdsi}",
|
||||||
|
]
|
||||||
|
|
||||||
|
else:
|
||||||
|
if td.getCodec() != TrackCodec.PGS:
|
||||||
|
inputMappingTokens += ["-map", f"0:{stdi}"]
|
||||||
|
|
||||||
|
return inputMappingTokens
|
||||||
|
|
||||||
|
|
||||||
|
def searchSubtitleFiles(self, searchDirectory, prefix):
|
||||||
|
|
||||||
|
sesl_match = re.compile(MediaDescriptor.SEASON_EPISODE_STREAM_LANGUAGE_MATCH)
|
||||||
|
|
||||||
|
subtitleFileDescriptors = []
|
||||||
|
for subtitleFilename in os.listdir(searchDirectory):
|
||||||
|
if subtitleFilename.startswith(prefix) and subtitleFilename.endswith(
|
||||||
|
"." + MediaDescriptor.SUBTITLE_FILE_EXTENSION
|
||||||
|
):
|
||||||
|
sesl_result = sesl_match.search(subtitleFilename)
|
||||||
|
if sesl_result is not None:
|
||||||
|
subtitleFilePath = os.path.join(searchDirectory, subtitleFilename)
|
||||||
|
if os.path.isfile(subtitleFilePath):
|
||||||
|
|
||||||
|
subtitleFileDescriptor = {}
|
||||||
|
subtitleFileDescriptor["path"] = subtitleFilePath
|
||||||
|
subtitleFileDescriptor["season"] = int(sesl_result.group(1))
|
||||||
|
subtitleFileDescriptor["episode"] = int(sesl_result.group(2))
|
||||||
|
subtitleFileDescriptor["index"] = int(sesl_result.group(3))
|
||||||
|
subtitleFileDescriptor["language"] = sesl_result.group(4)
|
||||||
|
|
||||||
|
dispSet = set()
|
||||||
|
dispCaptGroups = sesl_result.groups()
|
||||||
|
numCaptGroups = len(dispCaptGroups)
|
||||||
|
if numCaptGroups > 4:
|
||||||
|
for groupIndex in range(numCaptGroups - 4):
|
||||||
|
disp = TrackDisposition.fromIndicator(dispCaptGroups[groupIndex + 4])
|
||||||
|
if disp is not None:
|
||||||
|
dispSet.add(disp)
|
||||||
|
subtitleFileDescriptor["disposition_set"] = dispSet
|
||||||
|
|
||||||
|
subtitleFileDescriptors.append(subtitleFileDescriptor)
|
||||||
|
|
||||||
|
self.__logger.debug(f"searchSubtitleFiles(): Available subtitle files {subtitleFileDescriptors}")
|
||||||
|
|
||||||
|
return subtitleFileDescriptors
|
||||||
|
|
||||||
|
|
||||||
|
def importSubtitles(self, searchDirectory, prefix, season: int = -1, episode: int = -1):
|
||||||
|
|
||||||
|
# click.echo(f"Season: {season} Episode: {episode}")
|
||||||
|
self.__logger.debug(f"importSubtitles(): Season: {season} Episode: {episode}")
|
||||||
|
|
||||||
|
availableFileSubtitleDescriptors = self.searchSubtitleFiles(searchDirectory, prefix)
|
||||||
|
|
||||||
|
self.__logger.debug(f"importSubtitles(): availableFileSubtitleDescriptors: {availableFileSubtitleDescriptors}")
|
||||||
|
|
||||||
|
subtitleTracks = self.getSubtitleTracks()
|
||||||
|
|
||||||
|
self.__logger.debug(f"importSubtitles(): subtitleTracks: {[s.getIndex() for s in subtitleTracks]}")
|
||||||
|
|
||||||
|
matchingSubtitleFileDescriptors = (
|
||||||
|
sorted(
|
||||||
|
[
|
||||||
|
d
|
||||||
|
for d in availableFileSubtitleDescriptors
|
||||||
|
if d["season"] == int(season) and d["episode"] == int(episode)
|
||||||
|
],
|
||||||
|
key=lambda d: d["index"],
|
||||||
|
)
|
||||||
|
if availableFileSubtitleDescriptors
|
||||||
|
else []
|
||||||
|
)
|
||||||
|
|
||||||
|
self.__logger.debug(f"importSubtitles(): matchingSubtitleFileDescriptors: {matchingSubtitleFileDescriptors}")
|
||||||
|
|
||||||
|
for msfd in matchingSubtitleFileDescriptors:
|
||||||
|
matchingSubtitleTrackDescriptor = [s for s in subtitleTracks if s.getIndex() == msfd["index"]]
|
||||||
|
if matchingSubtitleTrackDescriptor:
|
||||||
|
# click.echo(f"Found matching subtitle file {msfd["path"]}\n")
|
||||||
|
self.__logger.debug(f"importSubtitles(): Found matching subtitle file {msfd['path']}")
|
||||||
|
matchingSubtitleTrackDescriptor[0].setExternalSourceFilePath(msfd["path"])
|
||||||
|
|
||||||
|
# TODO: Check if useful
|
||||||
|
# matchingSubtitleTrackDescriptor[0].setDispositionSet(msfd["disposition_set"])
|
||||||
|
|
||||||
|
|
||||||
|
def getConfiguration(self, label: str = ''):
|
||||||
|
yield f"--- {label if label else 'MediaDescriptor '+str(id(self))} {' '.join([str(k)+'='+str(v) for k,v in self.__mediaTags.items()])}"
|
||||||
|
for td in self.getAllTrackDescriptors():
|
||||||
|
yield (f"{td.getIndex()}:{td.getType().indicator()}:{td.getSubIndex()} "
|
||||||
|
+ '|'.join([d.indicator() for d in td.getDispositionSet()])
|
||||||
|
+ ' ' + ' '.join([str(k)+'='+str(v) for k,v in td.getTags().items()]))
|
||||||
684
src/ffx/media_details_screen.py
Normal file
684
src/ffx/media_details_screen.py
Normal file
@@ -0,0 +1,684 @@
|
|||||||
|
import os, click, re
|
||||||
|
|
||||||
|
from textual.screen import Screen
|
||||||
|
from textual.widgets import Header, Footer, Static, Button, Input, DataTable
|
||||||
|
from textual.containers import Grid
|
||||||
|
|
||||||
|
from ffx.audio_layout import AudioLayout
|
||||||
|
|
||||||
|
from .pattern_controller import PatternController
|
||||||
|
from .show_controller import ShowController
|
||||||
|
from .track_controller import TrackController
|
||||||
|
from .tag_controller import TagController
|
||||||
|
|
||||||
|
from .show_details_screen import ShowDetailsScreen
|
||||||
|
from .pattern_details_screen import PatternDetailsScreen
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
from ffx.track_codec import TrackCodec
|
||||||
|
from ffx.model.track import Track
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.show_descriptor import ShowDescriptor
|
||||||
|
|
||||||
|
from textual.widgets._data_table import CellDoesNotExist
|
||||||
|
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
from ffx.file_properties import FileProperties
|
||||||
|
|
||||||
|
from ffx.helper import DIFF_ADDED_KEY, DIFF_CHANGED_KEY, DIFF_REMOVED_KEY
|
||||||
|
|
||||||
|
|
||||||
|
# Screen[dict[int, str, int]]
|
||||||
|
class MediaDetailsScreen(Screen):
|
||||||
|
|
||||||
|
CSS = """
|
||||||
|
|
||||||
|
Grid {
|
||||||
|
grid-size: 5 8;
|
||||||
|
grid-rows: 8 2 2 2 2 8 2 2 8;
|
||||||
|
grid-columns: 25 25 120 10 75;
|
||||||
|
height: 100%;
|
||||||
|
width: 100%;
|
||||||
|
padding: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable .datatable--cursor {
|
||||||
|
background: darkorange;
|
||||||
|
color: black;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable .datatable--header {
|
||||||
|
background: steelblue;
|
||||||
|
color: white;
|
||||||
|
}
|
||||||
|
|
||||||
|
Input {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
Button {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable {
|
||||||
|
min-height: 40;
|
||||||
|
}
|
||||||
|
|
||||||
|
#toplabel {
|
||||||
|
height: 1;
|
||||||
|
}
|
||||||
|
.two {
|
||||||
|
column-span: 2;
|
||||||
|
}
|
||||||
|
.three {
|
||||||
|
column-span: 3;
|
||||||
|
}
|
||||||
|
|
||||||
|
.four {
|
||||||
|
column-span: 4;
|
||||||
|
}
|
||||||
|
.five {
|
||||||
|
column-span: 5;
|
||||||
|
}
|
||||||
|
|
||||||
|
.triple {
|
||||||
|
row-span: 3;
|
||||||
|
}
|
||||||
|
|
||||||
|
.box {
|
||||||
|
height: 100%;
|
||||||
|
border: solid green;
|
||||||
|
}
|
||||||
|
|
||||||
|
.purple {
|
||||||
|
tint: purple 40%;
|
||||||
|
}
|
||||||
|
|
||||||
|
.yellow {
|
||||||
|
tint: yellow 40%;
|
||||||
|
}
|
||||||
|
|
||||||
|
#differences-table {
|
||||||
|
row-span: 8;
|
||||||
|
/* tint: magenta 40%; */
|
||||||
|
}
|
||||||
|
|
||||||
|
/* #pattern_input {
|
||||||
|
tint: red 40%;
|
||||||
|
}*/
|
||||||
|
"""
|
||||||
|
|
||||||
|
|
||||||
|
BINDINGS = [
|
||||||
|
("n", "new_pattern", "New Pattern"),
|
||||||
|
("u", "update_pattern", "Update Pattern"),
|
||||||
|
("e", "edit_pattern", "Edit Pattern"),
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
super().__init__()
|
||||||
|
|
||||||
|
self.context = self.app.getContext()
|
||||||
|
self.Session = self.context['database']['session'] # convenience
|
||||||
|
|
||||||
|
self.__pc = PatternController(context = self.context)
|
||||||
|
self.__sc = ShowController(context = self.context)
|
||||||
|
self.__tc = TrackController(context = self.context)
|
||||||
|
self.__tac = TagController(context = self.context)
|
||||||
|
|
||||||
|
if not 'command' in self.context.keys() or self.context['command'] != 'inspect':
|
||||||
|
raise click.ClickException(f"MediaDetailsScreen.__init__(): Can only perform command 'inspect'")
|
||||||
|
|
||||||
|
if not 'arguments' in self.context.keys() or not 'filename' in self.context['arguments'].keys() or not self.context['arguments']['filename']:
|
||||||
|
raise click.ClickException(f"MediaDetailsScreen.__init__(): Argument 'filename' is required to be provided for command 'inspect'")
|
||||||
|
|
||||||
|
self.__mediaFilename = self.context['arguments']['filename']
|
||||||
|
|
||||||
|
if not os.path.isfile(self.__mediaFilename):
|
||||||
|
raise click.ClickException(f"MediaDetailsScreen.__init__(): Media file {self.__mediaFilename} does not exist")
|
||||||
|
|
||||||
|
self.loadProperties()
|
||||||
|
|
||||||
|
|
||||||
|
def removeShow(self, showId : int = -1):
|
||||||
|
"""Remove show entry from DataTable.
|
||||||
|
Removes the <New show> entry if showId is not set"""
|
||||||
|
|
||||||
|
for rowKey, row in self.showsTable.rows.items(): # dict[RowKey, Row]
|
||||||
|
|
||||||
|
rowData = self.showsTable.get_row(rowKey)
|
||||||
|
|
||||||
|
try:
|
||||||
|
if (showId == -1 and rowData[0] == ' '
|
||||||
|
or showId == int(rowData[0])):
|
||||||
|
self.showsTable.remove_row(rowKey)
|
||||||
|
return
|
||||||
|
except:
|
||||||
|
continue
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def getRowIndexFromShowId(self, showId : int = -1) -> int:
|
||||||
|
"""Find the index of the row where the value in the specified column matches the target_value."""
|
||||||
|
|
||||||
|
for rowKey, row in self.showsTable.rows.items(): # dict[RowKey, Row]
|
||||||
|
|
||||||
|
rowData = self.showsTable.get_row(rowKey)
|
||||||
|
|
||||||
|
try:
|
||||||
|
if ((showId == -1 and rowData[0] == ' ')
|
||||||
|
or showId == int(rowData[0])):
|
||||||
|
return int(self.showsTable.get_row_index(rowKey))
|
||||||
|
except:
|
||||||
|
continue
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def loadProperties(self):
|
||||||
|
|
||||||
|
self.__mediaFileProperties = FileProperties(self.context, self.__mediaFilename)
|
||||||
|
self.__currentMediaDescriptor = self.__mediaFileProperties.getMediaDescriptor()
|
||||||
|
|
||||||
|
#HINT: This is None if the filename did not match anything in database
|
||||||
|
self.__currentPattern = self.__mediaFileProperties.getPattern()
|
||||||
|
|
||||||
|
# keine tags vorhanden
|
||||||
|
self.__targetMediaDescriptor = self.__currentPattern.getMediaDescriptor(self.context) if self.__currentPattern is not None else None
|
||||||
|
|
||||||
|
# Enumerating differences between media descriptors
|
||||||
|
# from file (=current) vs from stored in database (=target)
|
||||||
|
try:
|
||||||
|
self.__mediaDifferences = self.__targetMediaDescriptor.compare(self.__currentMediaDescriptor) if self.__currentPattern is not None else {}
|
||||||
|
except ValueError:
|
||||||
|
self.__mediaDifferences = {}
|
||||||
|
|
||||||
|
|
||||||
|
def updateDifferences(self):
|
||||||
|
|
||||||
|
self.loadProperties()
|
||||||
|
|
||||||
|
self.differencesTable.clear()
|
||||||
|
|
||||||
|
if MediaDescriptor.TAGS_KEY in self.__mediaDifferences.keys():
|
||||||
|
|
||||||
|
currentTags = self.__currentMediaDescriptor.getTags()
|
||||||
|
targetTags = self.__targetMediaDescriptor.getTags()
|
||||||
|
|
||||||
|
if DIFF_ADDED_KEY in self.__mediaDifferences[MediaDescriptor.TAGS_KEY].keys():
|
||||||
|
for addedTagKey in self.__mediaDifferences[MediaDescriptor.TAGS_KEY][DIFF_ADDED_KEY]:
|
||||||
|
row = (f"added media tag: key='{addedTagKey}' value='{targetTags[addedTagKey]}'",)
|
||||||
|
self.differencesTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
if DIFF_REMOVED_KEY in self.__mediaDifferences[MediaDescriptor.TAGS_KEY].keys():
|
||||||
|
for removedTagKey in self.__mediaDifferences[MediaDescriptor.TAGS_KEY][DIFF_REMOVED_KEY]:
|
||||||
|
row = (f"removed media tag: key='{removedTagKey}' value='{currentTags[removedTagKey]}'",)
|
||||||
|
self.differencesTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
if DIFF_CHANGED_KEY in self.__mediaDifferences[MediaDescriptor.TAGS_KEY].keys():
|
||||||
|
for changedTagKey in self.__mediaDifferences[MediaDescriptor.TAGS_KEY][DIFF_CHANGED_KEY]:
|
||||||
|
row = (f"changed media tag: key='{changedTagKey}' value='{currentTags[changedTagKey]}'->'{targetTags[changedTagKey]}'",)
|
||||||
|
self.differencesTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
if MediaDescriptor.TRACKS_KEY in self.__mediaDifferences.keys():
|
||||||
|
|
||||||
|
currentTracks = self.__currentMediaDescriptor.getAllTrackDescriptors() # 0,1,2,3
|
||||||
|
targetTracks = self.__targetMediaDescriptor.getAllTrackDescriptors() # 0 <- from DB
|
||||||
|
|
||||||
|
if DIFF_ADDED_KEY in self.__mediaDifferences[MediaDescriptor.TRACKS_KEY].keys():
|
||||||
|
|
||||||
|
#raise click.ClickException(f"add track {self.__mediaDifferences[MediaDescriptor.TRACKS_KEY][DIFF_ADDED_KEY]}")
|
||||||
|
for addedTrackIndex in self.__mediaDifferences[MediaDescriptor.TRACKS_KEY][DIFF_ADDED_KEY]:
|
||||||
|
addedTrack : Track = targetTracks[addedTrackIndex]
|
||||||
|
row = (f"added {addedTrack.getType().label()} track: index={addedTrackIndex} lang={addedTrack.getLanguage().threeLetter()}",)
|
||||||
|
self.differencesTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
if DIFF_REMOVED_KEY in self.__mediaDifferences[MediaDescriptor.TRACKS_KEY].keys():
|
||||||
|
for removedTrackIndex in self.__mediaDifferences[MediaDescriptor.TRACKS_KEY][DIFF_REMOVED_KEY]:
|
||||||
|
row = (f"removed track: index={removedTrackIndex}",)
|
||||||
|
self.differencesTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
if DIFF_CHANGED_KEY in self.__mediaDifferences[MediaDescriptor.TRACKS_KEY].keys():
|
||||||
|
for changedTrackIndex in self.__mediaDifferences[MediaDescriptor.TRACKS_KEY][DIFF_CHANGED_KEY].keys():
|
||||||
|
|
||||||
|
changedTrack : Track = targetTracks[changedTrackIndex]
|
||||||
|
changedTrackDiff : dict = self.__mediaDifferences[MediaDescriptor.TRACKS_KEY][DIFF_CHANGED_KEY][changedTrackIndex]
|
||||||
|
|
||||||
|
if MediaDescriptor.TAGS_KEY in changedTrackDiff.keys():
|
||||||
|
|
||||||
|
if DIFF_ADDED_KEY in changedTrackDiff[MediaDescriptor.TAGS_KEY]:
|
||||||
|
for addedTagKey in changedTrackDiff[MediaDescriptor.TAGS_KEY][DIFF_ADDED_KEY]:
|
||||||
|
addedTagValue = changedTrack.getTags()[addedTagKey]
|
||||||
|
row = (f"changed {changedTrack.getType().label()} track index={changedTrackIndex} added key={addedTagKey} value={addedTagValue}",)
|
||||||
|
self.differencesTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
if DIFF_REMOVED_KEY in changedTrackDiff[MediaDescriptor.TAGS_KEY]:
|
||||||
|
for removedTagKey in changedTrackDiff[MediaDescriptor.TAGS_KEY][DIFF_REMOVED_KEY]:
|
||||||
|
row = (f"changed {changedTrack.getType().label()} track index={changedTrackIndex} removed key={removedTagKey}",)
|
||||||
|
self.differencesTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
if TrackDescriptor.DISPOSITION_SET_KEY in changedTrackDiff.keys():
|
||||||
|
|
||||||
|
if DIFF_ADDED_KEY in changedTrackDiff[TrackDescriptor.DISPOSITION_SET_KEY]:
|
||||||
|
for addedDisposition in changedTrackDiff[TrackDescriptor.DISPOSITION_SET_KEY][DIFF_ADDED_KEY]:
|
||||||
|
row = (f"changed {changedTrack.getType().label()} track index={changedTrackIndex} added disposition={addedDisposition.label()}",)
|
||||||
|
self.differencesTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
if DIFF_REMOVED_KEY in changedTrackDiff[TrackDescriptor.DISPOSITION_SET_KEY]:
|
||||||
|
for removedDisposition in changedTrackDiff[TrackDescriptor.DISPOSITION_SET_KEY][DIFF_REMOVED_KEY]:
|
||||||
|
row = (f"changed {changedTrack.getType().label()} track index={changedTrackIndex} removed disposition={removedDisposition.label()}",)
|
||||||
|
self.differencesTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
|
||||||
|
def on_mount(self):
|
||||||
|
|
||||||
|
if self.__currentPattern is None:
|
||||||
|
row = (' ', '<New show>', ' ') # Convert each element to a string before adding
|
||||||
|
self.showsTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
for show in self.__sc.getAllShows():
|
||||||
|
row = (int(show.id), show.name, show.year) # Convert each element to a string before adding
|
||||||
|
self.showsTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
for mediaTagKey, mediaTagValue in self.__currentMediaDescriptor.getTags().items():
|
||||||
|
row = (mediaTagKey, mediaTagValue) # Convert each element to a string before adding
|
||||||
|
self.mediaTagsTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
self.updateTracks()
|
||||||
|
|
||||||
|
|
||||||
|
if self.__currentPattern is not None:
|
||||||
|
|
||||||
|
showIdentifier = self.__currentPattern.getShowId()
|
||||||
|
showRowIndex = self.getRowIndexFromShowId(showIdentifier)
|
||||||
|
if showRowIndex is not None:
|
||||||
|
self.showsTable.move_cursor(row=showRowIndex)
|
||||||
|
|
||||||
|
self.query_one("#pattern_input", Input).value = self.__currentPattern.getPattern()
|
||||||
|
|
||||||
|
self.updateDifferences()
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
self.query_one("#pattern_input", Input).value = self.__mediaFilename
|
||||||
|
self.highlightPattern(True)
|
||||||
|
|
||||||
|
|
||||||
|
def highlightPattern(self, state : bool):
|
||||||
|
if state:
|
||||||
|
self.query_one("#pattern_input", Input).styles.background = 'red'
|
||||||
|
else:
|
||||||
|
self.query_one("#pattern_input", Input).styles.background = None
|
||||||
|
|
||||||
|
|
||||||
|
def updateTracks(self):
|
||||||
|
|
||||||
|
self.tracksTable.clear()
|
||||||
|
|
||||||
|
trackDescriptorList = self.__currentMediaDescriptor.getAllTrackDescriptors()
|
||||||
|
|
||||||
|
typeCounter = {}
|
||||||
|
|
||||||
|
for td in trackDescriptorList:
|
||||||
|
|
||||||
|
trackType = td.getType()
|
||||||
|
if not trackType in typeCounter.keys():
|
||||||
|
typeCounter[trackType] = 0
|
||||||
|
|
||||||
|
dispoSet = td.getDispositionSet()
|
||||||
|
audioLayout = td.getAudioLayout()
|
||||||
|
row = (td.getIndex(),
|
||||||
|
trackType.label(),
|
||||||
|
typeCounter[trackType],
|
||||||
|
td.getCodec().label(),
|
||||||
|
audioLayout.label() if trackType == TrackType.AUDIO
|
||||||
|
and audioLayout != AudioLayout.LAYOUT_UNDEFINED else ' ',
|
||||||
|
td.getLanguage().label(),
|
||||||
|
td.getTitle(),
|
||||||
|
'Yes' if TrackDisposition.DEFAULT in dispoSet else 'No',
|
||||||
|
'Yes' if TrackDisposition.FORCED in dispoSet else 'No')
|
||||||
|
|
||||||
|
self.tracksTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
typeCounter[trackType] += 1
|
||||||
|
|
||||||
|
|
||||||
|
def compose(self):
|
||||||
|
|
||||||
|
# Create the DataTable widget
|
||||||
|
self.showsTable = DataTable(classes="two")
|
||||||
|
|
||||||
|
# Define the columns with headers
|
||||||
|
self.column_key_show_id = self.showsTable.add_column("ID", width=10)
|
||||||
|
self.column_key_show_name = self.showsTable.add_column("Name", width=50)
|
||||||
|
self.column_key_show_year = self.showsTable.add_column("Year", width=10)
|
||||||
|
|
||||||
|
self.showsTable.cursor_type = 'row'
|
||||||
|
|
||||||
|
|
||||||
|
self.mediaTagsTable = DataTable(classes="two")
|
||||||
|
|
||||||
|
# Define the columns with headers
|
||||||
|
self.column_key_track_tag_key = self.mediaTagsTable.add_column("Key", width=50)
|
||||||
|
self.column_key_track_tag_value = self.mediaTagsTable.add_column("Value", width=100)
|
||||||
|
|
||||||
|
self.mediaTagsTable.cursor_type = 'row'
|
||||||
|
|
||||||
|
|
||||||
|
self.tracksTable = DataTable(classes="two")
|
||||||
|
|
||||||
|
# Define the columns with headers
|
||||||
|
self.column_key_track_index = self.tracksTable.add_column("Index", width=5)
|
||||||
|
self.column_key_track_type = self.tracksTable.add_column("Type", width=10)
|
||||||
|
self.column_key_track_sub_index = self.tracksTable.add_column("SubIndex", width=8)
|
||||||
|
self.column_key_track_codec = self.tracksTable.add_column("Codec", width=10)
|
||||||
|
self.column_key_track_layout = self.tracksTable.add_column("Layout", width=10)
|
||||||
|
self.column_key_track_language = self.tracksTable.add_column("Language", width=15)
|
||||||
|
self.column_key_track_title = self.tracksTable.add_column("Title", width=48)
|
||||||
|
self.column_key_track_default = self.tracksTable.add_column("Default", width=8)
|
||||||
|
self.column_key_track_forced = self.tracksTable.add_column("Forced", width=8)
|
||||||
|
|
||||||
|
self.tracksTable.cursor_type = 'row'
|
||||||
|
|
||||||
|
|
||||||
|
# Create the DataTable widget
|
||||||
|
self.differencesTable = DataTable(id='differences-table') # classes="triple"
|
||||||
|
|
||||||
|
# Define the columns with headers
|
||||||
|
self.column_key_differences = self.differencesTable.add_column("Differences (file->db)", width=70)
|
||||||
|
|
||||||
|
self.differencesTable.cursor_type = 'row'
|
||||||
|
|
||||||
|
yield Header()
|
||||||
|
|
||||||
|
with Grid():
|
||||||
|
|
||||||
|
# 1
|
||||||
|
yield Static("Show")
|
||||||
|
yield self.showsTable
|
||||||
|
yield Static(" ")
|
||||||
|
yield self.differencesTable
|
||||||
|
|
||||||
|
# 2
|
||||||
|
yield Static(" ", classes="four")
|
||||||
|
|
||||||
|
# 3
|
||||||
|
yield Static(" ")
|
||||||
|
yield Button("Substitute", id="pattern_button")
|
||||||
|
yield Static(" ", classes="two")
|
||||||
|
|
||||||
|
# 4
|
||||||
|
yield Static("Pattern")
|
||||||
|
yield Input(type="text", id='pattern_input', classes="two")
|
||||||
|
|
||||||
|
yield Static(" ")
|
||||||
|
|
||||||
|
# 5
|
||||||
|
yield Static(" ", classes="four")
|
||||||
|
|
||||||
|
# 6
|
||||||
|
yield Static("Media Tags")
|
||||||
|
yield self.mediaTagsTable
|
||||||
|
yield Static(" ")
|
||||||
|
|
||||||
|
# 7
|
||||||
|
yield Static(" ", classes="four")
|
||||||
|
|
||||||
|
# 8
|
||||||
|
yield Static(" ")
|
||||||
|
yield Button("Set Default", id="select_default_button")
|
||||||
|
yield Button("Set Forced", id="select_forced_button")
|
||||||
|
yield Static(" ")
|
||||||
|
# 9
|
||||||
|
yield Static("Streams")
|
||||||
|
yield self.tracksTable
|
||||||
|
yield Static(" ")
|
||||||
|
|
||||||
|
yield Footer()
|
||||||
|
|
||||||
|
|
||||||
|
def getPatternDescriptorFromInput(self):
|
||||||
|
"""Returns show id and pattern from corresponding inputs"""
|
||||||
|
patternDescriptor = {}
|
||||||
|
try:
|
||||||
|
patternDescriptor['show_id'] = self.getSelectedShowDescriptor().getId()
|
||||||
|
patternDescriptor['pattern'] = str(self.query_one("#pattern_input", Input).value)
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
return patternDescriptor
|
||||||
|
|
||||||
|
|
||||||
|
def on_button_pressed(self, event: Button.Pressed) -> None:
|
||||||
|
|
||||||
|
if event.button.id == "pattern_button":
|
||||||
|
|
||||||
|
pattern = self.query_one("#pattern_input", Input).value
|
||||||
|
|
||||||
|
patternMatch = re.search(FileProperties.SE_INDICATOR_PATTERN, pattern)
|
||||||
|
|
||||||
|
if patternMatch:
|
||||||
|
self.query_one("#pattern_input", Input).value = pattern.replace(patternMatch.group(1), FileProperties.SE_INDICATOR_PATTERN)
|
||||||
|
|
||||||
|
|
||||||
|
if event.button.id == "select_default_button":
|
||||||
|
selectedTrackDescriptor = self.getSelectedTrackDescriptor()
|
||||||
|
self.__currentMediaDescriptor.setDefaultSubTrack(selectedTrackDescriptor.getType(), selectedTrackDescriptor.getSubIndex())
|
||||||
|
self.updateTracks()
|
||||||
|
|
||||||
|
if event.button.id == "select_forced_button":
|
||||||
|
selectedTrackDescriptor = self.getSelectedTrackDescriptor()
|
||||||
|
self.__currentMediaDescriptor.setForcedSubTrack(selectedTrackDescriptor.getType(), selectedTrackDescriptor.getSubIndex())
|
||||||
|
self.updateTracks()
|
||||||
|
|
||||||
|
|
||||||
|
def getSelectedTrackDescriptor(self):
|
||||||
|
"""Returns a partial track descriptor"""
|
||||||
|
try:
|
||||||
|
|
||||||
|
# Fetch the currently selected row when 'Enter' is pressed
|
||||||
|
#selected_row_index = self.table.cursor_row
|
||||||
|
row_key, col_key = self.tracksTable.coordinate_to_cell_key(self.tracksTable.cursor_coordinate)
|
||||||
|
|
||||||
|
if row_key is not None:
|
||||||
|
selected_track_data = self.tracksTable.get_row(row_key)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self.context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = int(selected_track_data[0])
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.fromLabel(selected_track_data[1])
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = int(selected_track_data[2])
|
||||||
|
kwargs[TrackDescriptor.CODEC_KEY] = TrackCodec.fromLabel(selected_track_data[3])
|
||||||
|
kwargs[TrackDescriptor.AUDIO_LAYOUT_KEY] = AudioLayout.fromLabel(selected_track_data[4])
|
||||||
|
|
||||||
|
return TrackDescriptor(**kwargs)
|
||||||
|
else:
|
||||||
|
return None
|
||||||
|
|
||||||
|
except CellDoesNotExist:
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def getSelectedShowDescriptor(self) -> ShowDescriptor:
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
row_key, col_key = self.showsTable.coordinate_to_cell_key(self.showsTable.cursor_coordinate)
|
||||||
|
|
||||||
|
if row_key is not None:
|
||||||
|
selected_row_data = self.showsTable.get_row(row_key)
|
||||||
|
|
||||||
|
try:
|
||||||
|
kwargs = {}
|
||||||
|
|
||||||
|
kwargs[ShowDescriptor.ID_KEY] = int(selected_row_data[0])
|
||||||
|
kwargs[ShowDescriptor.NAME_KEY] = str(selected_row_data[1])
|
||||||
|
kwargs[ShowDescriptor.YEAR_KEY] = int(selected_row_data[2])
|
||||||
|
|
||||||
|
return ShowDescriptor(**kwargs)
|
||||||
|
|
||||||
|
except ValueError:
|
||||||
|
return None
|
||||||
|
|
||||||
|
except CellDoesNotExist:
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def handle_new_pattern(self, showDescriptor: ShowDescriptor):
|
||||||
|
|
||||||
|
if type(showDescriptor) is not ShowDescriptor:
|
||||||
|
raise TypeError("MediaDetailsScreen.handle_new_pattern(): Argument 'showDescriptor' has to be of type ShowDescriptor")
|
||||||
|
|
||||||
|
showRowIndex = self.getRowIndexFromShowId(showDescriptor.getId())
|
||||||
|
if showRowIndex is None:
|
||||||
|
show = (showDescriptor.getId(), showDescriptor.getName(), showDescriptor.getYear())
|
||||||
|
self.showsTable.add_row(*map(str, show))
|
||||||
|
|
||||||
|
showRowIndex = self.getRowIndexFromShowId(showDescriptor.getId())
|
||||||
|
if showRowIndex is not None:
|
||||||
|
self.showsTable.move_cursor(row=showRowIndex)
|
||||||
|
|
||||||
|
self.removeShow()
|
||||||
|
|
||||||
|
patternDescriptor = self.getPatternDescriptorFromInput()
|
||||||
|
|
||||||
|
if patternDescriptor:
|
||||||
|
patternId = self.__pc.addPattern(patternDescriptor)
|
||||||
|
if patternId:
|
||||||
|
self.highlightPattern(False)
|
||||||
|
|
||||||
|
for tagKey, tagValue in self.__currentMediaDescriptor.getTags().items():
|
||||||
|
self.__tac.updateMediaTag(patternId, tagKey, tagValue)
|
||||||
|
|
||||||
|
for trackDescriptor in self.__currentMediaDescriptor.getAllTrackDescriptors():
|
||||||
|
self.__tc.addTrack(trackDescriptor, patternId = patternId)
|
||||||
|
|
||||||
|
|
||||||
|
def action_new_pattern(self):
|
||||||
|
|
||||||
|
#TODO #427: Fehlermeldung in TUI
|
||||||
|
# try:
|
||||||
|
# self.__currentMediaDescriptor.checkConfiguration()
|
||||||
|
# except ValueError:
|
||||||
|
# return
|
||||||
|
|
||||||
|
selectedShowDescriptor = self.getSelectedShowDescriptor()
|
||||||
|
|
||||||
|
#HINT: Callback is invoked after this method has exited. As a workaround the callback is executed directly
|
||||||
|
# from here with a mock-up screen result containing the necessary part of keys to perform correctly.
|
||||||
|
if selectedShowDescriptor is None:
|
||||||
|
self.app.push_screen(ShowDetailsScreen(), self.handle_new_pattern)
|
||||||
|
else:
|
||||||
|
self.handle_new_pattern(selectedShowDescriptor)
|
||||||
|
|
||||||
|
|
||||||
|
def action_update_pattern(self):
|
||||||
|
"""When updating the database the actions must reverse the difference (eq to diff db->file)"""
|
||||||
|
|
||||||
|
if self.__currentPattern is not None:
|
||||||
|
patternDescriptor = self.getPatternDescriptorFromInput()
|
||||||
|
if (patternDescriptor
|
||||||
|
and self.__currentPattern.getPattern() != patternDescriptor['pattern']):
|
||||||
|
return self.__pc.updatePattern(self.__currentPattern.getId(), patternDescriptor)
|
||||||
|
|
||||||
|
self.loadProperties()
|
||||||
|
|
||||||
|
if MediaDescriptor.TAGS_KEY in self.__mediaDifferences.keys():
|
||||||
|
|
||||||
|
if DIFF_ADDED_KEY in self.__mediaDifferences[MediaDescriptor.TAGS_KEY].keys():
|
||||||
|
for addedTagKey in self.__mediaDifferences[MediaDescriptor.TAGS_KEY][DIFF_ADDED_KEY]:
|
||||||
|
self.__tac.deleteMediaTagByKey(self.__currentPattern.getId(), addedTagKey)
|
||||||
|
|
||||||
|
if DIFF_REMOVED_KEY in self.__mediaDifferences[MediaDescriptor.TAGS_KEY].keys():
|
||||||
|
for removedTagKey in self.__mediaDifferences[MediaDescriptor.TAGS_KEY][DIFF_REMOVED_KEY]:
|
||||||
|
currentTags = self.__currentMediaDescriptor.getTags()
|
||||||
|
self.__tac.updateMediaTag(self.__currentPattern.getId(), removedTagKey, currentTags[removedTagKey])
|
||||||
|
|
||||||
|
if DIFF_CHANGED_KEY in self.__mediaDifferences[MediaDescriptor.TAGS_KEY].keys():
|
||||||
|
for changedTagKey in self.__mediaDifferences[MediaDescriptor.TAGS_KEY][DIFF_CHANGED_KEY]:
|
||||||
|
currentTags = self.__currentMediaDescriptor.getTags()
|
||||||
|
self.__tac.updateMediaTag(self.__currentPattern.getId(), changedTagKey, currentTags[changedTagKey])
|
||||||
|
|
||||||
|
if MediaDescriptor.TRACKS_KEY in self.__mediaDifferences.keys():
|
||||||
|
|
||||||
|
if DIFF_ADDED_KEY in self.__mediaDifferences[MediaDescriptor.TRACKS_KEY].keys():
|
||||||
|
|
||||||
|
for addedTrackIndex in self.__mediaDifferences[MediaDescriptor.TRACKS_KEY][DIFF_ADDED_KEY]:
|
||||||
|
targetTracks = [t for t in self.__targetMediaDescriptor.getAllTrackDescriptors() if t.getIndex() == addedTrackIndex]
|
||||||
|
if targetTracks:
|
||||||
|
self.__tc.deleteTrack(targetTracks[0].getId()) # id
|
||||||
|
|
||||||
|
if DIFF_REMOVED_KEY in self.__mediaDifferences[MediaDescriptor.TRACKS_KEY].keys():
|
||||||
|
for removedTrackIndex, removedTrack in self.__mediaDifferences[MediaDescriptor.TRACKS_KEY][DIFF_REMOVED_KEY].items():
|
||||||
|
|
||||||
|
# Track per inspect/update hinzufügen
|
||||||
|
self.__tc.addTrack(removedTrack, patternId = self.__currentPattern.getId())
|
||||||
|
|
||||||
|
if DIFF_CHANGED_KEY in self.__mediaDifferences[MediaDescriptor.TRACKS_KEY].keys():
|
||||||
|
|
||||||
|
# [vsTracks[tp].getIndex()] = trackDiff
|
||||||
|
for changedTrackIndex, changedTrackDiff in self.__mediaDifferences[MediaDescriptor.TRACKS_KEY][DIFF_CHANGED_KEY].items():
|
||||||
|
|
||||||
|
changedTargetTracks = [t for t in self.__targetMediaDescriptor.getAllTrackDescriptors() if t.getIndex() == changedTrackIndex]
|
||||||
|
changedTargeTrackId = changedTargetTracks[0].getId() if changedTargetTracks else None
|
||||||
|
changedTargetTrackIndex = changedTargetTracks[0].getIndex() if changedTargetTracks else None
|
||||||
|
|
||||||
|
changedCurrentTracks = [t for t in self.__currentMediaDescriptor.getAllTrackDescriptors() if t.getIndex() == changedTrackIndex]
|
||||||
|
# changedCurrentTrackId #HINT: Undefined as track descriptors do not come from file with track_id
|
||||||
|
|
||||||
|
if TrackDescriptor.TAGS_KEY in changedTrackDiff.keys():
|
||||||
|
changedTrackTagsDiff = changedTrackDiff[TrackDescriptor.TAGS_KEY]
|
||||||
|
|
||||||
|
if DIFF_ADDED_KEY in changedTrackTagsDiff.keys():
|
||||||
|
for addedTrackTagKey in changedTrackTagsDiff[DIFF_ADDED_KEY]:
|
||||||
|
|
||||||
|
if changedTargetTracks:
|
||||||
|
self.__tac.deleteTrackTagByKey(changedTargeTrackId, addedTrackTagKey)
|
||||||
|
|
||||||
|
if DIFF_REMOVED_KEY in changedTrackTagsDiff.keys():
|
||||||
|
for removedTrackTagKey in changedTrackTagsDiff[DIFF_REMOVED_KEY]:
|
||||||
|
if changedCurrentTracks:
|
||||||
|
self.__tac.updateTrackTag(changedTargeTrackId, removedTrackTagKey, changedCurrentTracks[0].getTags()[removedTrackTagKey])
|
||||||
|
|
||||||
|
if DIFF_CHANGED_KEY in changedTrackTagsDiff.keys():
|
||||||
|
for changedTrackTagKey in changedTrackTagsDiff[DIFF_CHANGED_KEY]:
|
||||||
|
if changedCurrentTracks:
|
||||||
|
self.__tac.updateTrackTag(changedTargeTrackId, changedTrackTagKey, changedCurrentTracks[0].getTags()[changedTrackTagKey])
|
||||||
|
|
||||||
|
if TrackDescriptor.DISPOSITION_SET_KEY in changedTrackDiff.keys():
|
||||||
|
changedTrackDispositionDiff = changedTrackDiff[TrackDescriptor.DISPOSITION_SET_KEY]
|
||||||
|
|
||||||
|
if DIFF_ADDED_KEY in changedTrackDispositionDiff.keys():
|
||||||
|
for changedTrackAddedDisposition in changedTrackDispositionDiff[DIFF_ADDED_KEY]:
|
||||||
|
if changedTargetTrackIndex is not None:
|
||||||
|
self.__tc.setDispositionState(self.__currentPattern.getId(), changedTargetTrackIndex, changedTrackAddedDisposition, False)
|
||||||
|
|
||||||
|
if DIFF_REMOVED_KEY in changedTrackDispositionDiff.keys():
|
||||||
|
for changedTrackRemovedDisposition in changedTrackDispositionDiff[DIFF_REMOVED_KEY]:
|
||||||
|
if changedTargetTrackIndex is not None:
|
||||||
|
self.__tc.setDispositionState(self.__currentPattern.getId(), changedTargetTrackIndex, changedTrackRemovedDisposition, True)
|
||||||
|
|
||||||
|
|
||||||
|
self.updateDifferences()
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def action_edit_pattern(self):
|
||||||
|
|
||||||
|
patternDescriptor = self.getPatternDescriptorFromInput()
|
||||||
|
|
||||||
|
if patternDescriptor['pattern']:
|
||||||
|
|
||||||
|
selectedPatternId = self.__pc.findPattern(patternDescriptor)
|
||||||
|
|
||||||
|
if selectedPatternId is None:
|
||||||
|
raise click.ClickException(f"MediaDetailsScreen.action_edit_pattern(): Pattern to edit has no id")
|
||||||
|
|
||||||
|
self.app.push_screen(PatternDetailsScreen(patternId = selectedPatternId, showId = self.getSelectedShowDescriptor().getId()), self.handle_edit_pattern) # <-
|
||||||
|
|
||||||
|
|
||||||
|
def handle_edit_pattern(self, screenResult):
|
||||||
|
self.query_one("#pattern_input", Input).value = screenResult['pattern']
|
||||||
|
self.updateDifferences()
|
||||||
|
|
||||||
0
src/ffx/model/__init__.py
Normal file
0
src/ffx/model/__init__.py
Normal file
0
src/ffx/model/conversions/__init__.py
Normal file
0
src/ffx/model/conversions/__init__.py
Normal file
47
src/ffx/model/conversions/conversion.py
Normal file
47
src/ffx/model/conversions/conversion.py
Normal file
@@ -0,0 +1,47 @@
|
|||||||
|
import os, sys, importlib, inspect, glob, re
|
||||||
|
|
||||||
|
from ffx.configuration_controller import ConfigurationController
|
||||||
|
from ffx.database import databaseContext
|
||||||
|
|
||||||
|
from sqlalchemy import Engine
|
||||||
|
from sqlalchemy.orm import sessionmaker
|
||||||
|
|
||||||
|
|
||||||
|
class Conversion():
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
|
||||||
|
self._context = {}
|
||||||
|
self._context['config'] = ConfigurationController()
|
||||||
|
|
||||||
|
self._context['database'] = databaseContext(databasePath=self._context['config'].getDatabaseFilePath())
|
||||||
|
|
||||||
|
self.__databaseSession: sessionmaker = self._context['database']['session']
|
||||||
|
self.__databaseEngine: Engine = self._context['database']['engine']
|
||||||
|
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def list():
|
||||||
|
|
||||||
|
basePath = os.path.dirname(__file__)
|
||||||
|
|
||||||
|
filenamePattern = re.compile("conversion_([0-9]+)_([0-9]+)\\.py")
|
||||||
|
|
||||||
|
filenameList = [os.path.basename(fp) for fp in glob.glob(f"{ basePath }/*.py") if fp != __file__]
|
||||||
|
|
||||||
|
versionTupleList = [(fm.group(1), fm.group(2)) for fn in filenameList if (fm := filenamePattern.search(fn))]
|
||||||
|
|
||||||
|
return versionTupleList
|
||||||
|
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getClassReference(versionFrom, versionTo):
|
||||||
|
importlib.import_module(f"ffx.model.conversions.conversion_{ versionFrom }_{ versionTo }")
|
||||||
|
for name, obj in inspect.getmembers(sys.modules[f"ffx.model.conversions.conversion_{ versionFrom }_{ versionTo }"]):
|
||||||
|
#HINT: Excluding DispositionCombination as it seems to be included by import (?)
|
||||||
|
if inspect.isclass(obj) and name != 'Conversion' and name.startswith('Conversion'):
|
||||||
|
return obj
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getAllClassReferences():
|
||||||
|
return [Conversion.getClassReference(verFrom, verTo) for verFrom, verTo in Conversion.list()]
|
||||||
17
src/ffx/model/conversions/conversion_2_3.py
Normal file
17
src/ffx/model/conversions/conversion_2_3.py
Normal file
@@ -0,0 +1,17 @@
|
|||||||
|
import os, sys, importlib, inspect, glob, re
|
||||||
|
|
||||||
|
from .conversion import Conversion
|
||||||
|
|
||||||
|
|
||||||
|
class Conversion_2_3(Conversion):
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
super().__init__()
|
||||||
|
|
||||||
|
def applyConversion(self):
|
||||||
|
|
||||||
|
s = self.__databaseSession()
|
||||||
|
e = self.__databaseEngine
|
||||||
|
|
||||||
|
with e.connect() as c:
|
||||||
|
c.execute("ALTER TABLE user ADD COLUMN email VARCHAR(255)")
|
||||||
7
src/ffx/model/conversions/conversion_3_4.py
Normal file
7
src/ffx/model/conversions/conversion_3_4.py
Normal file
@@ -0,0 +1,7 @@
|
|||||||
|
import os, sys, importlib, inspect, glob, re
|
||||||
|
|
||||||
|
from .conversion import Conversion
|
||||||
|
|
||||||
|
|
||||||
|
class Conversion_3_4(Conversion):
|
||||||
|
pass
|
||||||
28
src/ffx/model/media_tag.py
Normal file
28
src/ffx/model/media_tag.py
Normal file
@@ -0,0 +1,28 @@
|
|||||||
|
# from typing import List
|
||||||
|
from sqlalchemy import create_engine, Column, Integer, String, ForeignKey, Enum
|
||||||
|
from sqlalchemy.orm import relationship, declarative_base, sessionmaker
|
||||||
|
|
||||||
|
from .show import Base
|
||||||
|
|
||||||
|
|
||||||
|
class MediaTag(Base):
|
||||||
|
"""
|
||||||
|
relationship(argument, opt1, opt2, ...)
|
||||||
|
argument is string of class or Mapped class of the target entity
|
||||||
|
backref creates a bi-directional corresponding relationship (back_populates preferred)
|
||||||
|
back_populates points to the corresponding relationship (the actual class attribute identifier)
|
||||||
|
|
||||||
|
See: https://docs.sqlalchemy.org/en/(14|20)/orm/basic_relationships.html
|
||||||
|
"""
|
||||||
|
|
||||||
|
__tablename__ = 'media_tags'
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
id = Column(Integer, primary_key=True)
|
||||||
|
|
||||||
|
key = Column(String)
|
||||||
|
value = Column(String)
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
pattern_id = Column(Integer, ForeignKey('patterns.id', ondelete="CASCADE"))
|
||||||
|
pattern = relationship('Pattern', back_populates='media_tags')
|
||||||
76
src/ffx/model/pattern.py
Normal file
76
src/ffx/model/pattern.py
Normal file
@@ -0,0 +1,76 @@
|
|||||||
|
import click
|
||||||
|
|
||||||
|
from sqlalchemy import Column, Integer, String, ForeignKey
|
||||||
|
from sqlalchemy.orm import relationship
|
||||||
|
|
||||||
|
from .show import Base, Show
|
||||||
|
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
from ffx.show_descriptor import ShowDescriptor
|
||||||
|
|
||||||
|
|
||||||
|
class Pattern(Base):
|
||||||
|
|
||||||
|
__tablename__ = 'patterns'
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
id = Column(Integer, primary_key=True)
|
||||||
|
pattern = Column(String)
|
||||||
|
|
||||||
|
# v2.0
|
||||||
|
# id: Mapped[int] = mapped_column(Integer, primary_key=True)
|
||||||
|
# pattern: Mapped[str] = mapped_column(String, nullable=False)
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
show_id = Column(Integer, ForeignKey('shows.id', ondelete="CASCADE"))
|
||||||
|
show = relationship(Show, back_populates='patterns', lazy='joined')
|
||||||
|
|
||||||
|
# v2.0
|
||||||
|
# show_id: Mapped[int] = mapped_column(ForeignKey("shows.id", ondelete="CASCADE"))
|
||||||
|
# show: Mapped["Show"] = relationship(back_populates="patterns")
|
||||||
|
|
||||||
|
tracks = relationship('Track', back_populates='pattern', cascade="all, delete", lazy='joined')
|
||||||
|
|
||||||
|
|
||||||
|
media_tags = relationship('MediaTag', back_populates='pattern', cascade="all, delete", lazy='joined')
|
||||||
|
|
||||||
|
|
||||||
|
def getId(self):
|
||||||
|
return int(self.id)
|
||||||
|
|
||||||
|
def getShowId(self):
|
||||||
|
return int(self.show_id)
|
||||||
|
|
||||||
|
def getShowDescriptor(self, context) -> ShowDescriptor:
|
||||||
|
# click.echo(f"self.show {self.show} id={self.show_id}")
|
||||||
|
return self.show.getDescriptor(context)
|
||||||
|
|
||||||
|
def getId(self):
|
||||||
|
return int(self.id)
|
||||||
|
|
||||||
|
def getPattern(self):
|
||||||
|
return str(self.pattern)
|
||||||
|
|
||||||
|
def getTags(self):
|
||||||
|
return {str(t.key):str(t.value) for t in self.media_tags}
|
||||||
|
|
||||||
|
|
||||||
|
def getMediaDescriptor(self, context):
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
|
||||||
|
kwargs[MediaDescriptor.CONTEXT_KEY] = context
|
||||||
|
|
||||||
|
kwargs[MediaDescriptor.TAGS_KEY] = self.getTags()
|
||||||
|
kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY] = []
|
||||||
|
|
||||||
|
# Set ordered subindices
|
||||||
|
subIndexCounter = {}
|
||||||
|
for track in self.tracks:
|
||||||
|
trackType = track.getType()
|
||||||
|
if not trackType in subIndexCounter.keys():
|
||||||
|
subIndexCounter[trackType] = 0
|
||||||
|
kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY].append(track.getDescriptor(context, subIndex = subIndexCounter[trackType]))
|
||||||
|
subIndexCounter[trackType] += 1
|
||||||
|
|
||||||
|
return MediaDescriptor(**kwargs)
|
||||||
16
src/ffx/model/property.py
Normal file
16
src/ffx/model/property.py
Normal file
@@ -0,0 +1,16 @@
|
|||||||
|
# from typing import List
|
||||||
|
from sqlalchemy import create_engine, Column, Integer, String, ForeignKey, Enum
|
||||||
|
from sqlalchemy.orm import relationship, declarative_base, sessionmaker
|
||||||
|
|
||||||
|
from .show import Base
|
||||||
|
|
||||||
|
|
||||||
|
class Property(Base):
|
||||||
|
|
||||||
|
__tablename__ = 'properties'
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
id = Column(Integer, primary_key=True)
|
||||||
|
|
||||||
|
key = Column(String)
|
||||||
|
value = Column(String)
|
||||||
71
src/ffx/model/shifted_season.py
Normal file
71
src/ffx/model/shifted_season.py
Normal file
@@ -0,0 +1,71 @@
|
|||||||
|
import click
|
||||||
|
|
||||||
|
from sqlalchemy import Column, Integer, ForeignKey
|
||||||
|
from sqlalchemy.orm import relationship
|
||||||
|
|
||||||
|
from .show import Base, Show
|
||||||
|
|
||||||
|
|
||||||
|
class ShiftedSeason(Base):
|
||||||
|
|
||||||
|
__tablename__ = 'shifted_seasons'
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
id = Column(Integer, primary_key=True)
|
||||||
|
|
||||||
|
|
||||||
|
# v2.0
|
||||||
|
# id: Mapped[int] = mapped_column(Integer, primary_key=True)
|
||||||
|
# pattern: Mapped[str] = mapped_column(String, nullable=False)
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
show_id = Column(Integer, ForeignKey('shows.id', ondelete="CASCADE"))
|
||||||
|
show = relationship(Show, back_populates='shifted_seasons', lazy='joined')
|
||||||
|
|
||||||
|
# v2.0
|
||||||
|
# show_id: Mapped[int] = mapped_column(ForeignKey("shows.id", ondelete="CASCADE"))
|
||||||
|
# show: Mapped["Show"] = relationship(back_populates="patterns")
|
||||||
|
|
||||||
|
|
||||||
|
original_season = Column(Integer)
|
||||||
|
|
||||||
|
first_episode = Column(Integer, default = -1)
|
||||||
|
last_episode = Column(Integer, default = -1)
|
||||||
|
|
||||||
|
season_offset = Column(Integer, default = 0)
|
||||||
|
episode_offset = Column(Integer, default = 0)
|
||||||
|
|
||||||
|
|
||||||
|
def getId(self):
|
||||||
|
return self.id
|
||||||
|
|
||||||
|
|
||||||
|
def getOriginalSeason(self):
|
||||||
|
return self.original_season
|
||||||
|
|
||||||
|
def getFirstEpisode(self):
|
||||||
|
return self.first_episode
|
||||||
|
|
||||||
|
def getLastEpisode(self):
|
||||||
|
return self.last_episode
|
||||||
|
|
||||||
|
|
||||||
|
def getSeasonOffset(self):
|
||||||
|
return self.season_offset
|
||||||
|
|
||||||
|
def getEpisodeOffset(self):
|
||||||
|
return self.episode_offset
|
||||||
|
|
||||||
|
|
||||||
|
def getObj(self):
|
||||||
|
|
||||||
|
shiftedSeasonObj = {}
|
||||||
|
|
||||||
|
shiftedSeasonObj['original_season'] = self.getOriginalSeason()
|
||||||
|
shiftedSeasonObj['first_episode'] = self.getFirstEpisode()
|
||||||
|
shiftedSeasonObj['last_episode'] = self.getLastEpisode()
|
||||||
|
shiftedSeasonObj['season_offset'] = self.getSeasonOffset()
|
||||||
|
shiftedSeasonObj['episode_offset'] = self.getEpisodeOffset()
|
||||||
|
|
||||||
|
return shiftedSeasonObj
|
||||||
|
|
||||||
62
src/ffx/model/show.py
Normal file
62
src/ffx/model/show.py
Normal file
@@ -0,0 +1,62 @@
|
|||||||
|
# from typing import List
|
||||||
|
from sqlalchemy import create_engine, Column, Integer, String, ForeignKey
|
||||||
|
from sqlalchemy.orm import relationship, declarative_base, sessionmaker
|
||||||
|
|
||||||
|
from ffx.show_descriptor import ShowDescriptor
|
||||||
|
|
||||||
|
Base = declarative_base()
|
||||||
|
|
||||||
|
|
||||||
|
class Show(Base):
|
||||||
|
"""
|
||||||
|
relationship(argument, opt1, opt2, ...)
|
||||||
|
argument is string of class or Mapped class of the target entity
|
||||||
|
backref creates a bi-directional corresponding relationship (back_populates preferred)
|
||||||
|
back_populates points to the corresponding relationship (the actual class attribute identifier)
|
||||||
|
|
||||||
|
See: https://docs.sqlalchemy.org/en/(14|20)/orm/basic_relationships.html
|
||||||
|
"""
|
||||||
|
|
||||||
|
__tablename__ = 'shows'
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
id = Column(Integer, primary_key=True)
|
||||||
|
|
||||||
|
name = Column(String)
|
||||||
|
year = Column(Integer)
|
||||||
|
|
||||||
|
# v2.0
|
||||||
|
# id: Mapped[int] = mapped_column(Integer, primary_key=True)
|
||||||
|
# name: Mapped[str] = mapped_column(String, nullable=False)
|
||||||
|
# year: Mapped[int] = mapped_column(Integer, nullable=False)
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
#patterns = relationship('Pattern', back_populates='show', cascade="all, delete", passive_deletes=True)
|
||||||
|
patterns = relationship('Pattern', back_populates='show', cascade="all, delete")
|
||||||
|
# patterns = relationship('Pattern', back_populates='show', cascade="all")
|
||||||
|
|
||||||
|
# v2.0
|
||||||
|
# patterns: Mapped[List["Pattern"]] = relationship(back_populates="show", cascade="all, delete")
|
||||||
|
|
||||||
|
shifted_seasons = relationship('ShiftedSeason', back_populates='show', cascade="all, delete")
|
||||||
|
|
||||||
|
|
||||||
|
index_season_digits = Column(Integer, default=ShowDescriptor.DEFAULT_INDEX_SEASON_DIGITS)
|
||||||
|
index_episode_digits = Column(Integer, default=ShowDescriptor.DEFAULT_INDEX_EPISODE_DIGITS)
|
||||||
|
indicator_season_digits = Column(Integer, default=ShowDescriptor.DEFAULT_INDICATOR_SEASON_DIGITS)
|
||||||
|
indicator_episode_digits = Column(Integer, default=ShowDescriptor.DEFAULT_INDICATOR_EPISODE_DIGITS)
|
||||||
|
|
||||||
|
|
||||||
|
def getDescriptor(self, context):
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[ShowDescriptor.CONTEXT_KEY] = context
|
||||||
|
kwargs[ShowDescriptor.ID_KEY] = int(self.id)
|
||||||
|
kwargs[ShowDescriptor.NAME_KEY] = str(self.name)
|
||||||
|
kwargs[ShowDescriptor.YEAR_KEY] = int(self.year)
|
||||||
|
kwargs[ShowDescriptor.INDEX_SEASON_DIGITS_KEY] = int(self.index_season_digits)
|
||||||
|
kwargs[ShowDescriptor.INDEX_EPISODE_DIGITS_KEY] = int(self.index_episode_digits)
|
||||||
|
kwargs[ShowDescriptor.INDICATOR_SEASON_DIGITS_KEY] = int(self.indicator_season_digits)
|
||||||
|
kwargs[ShowDescriptor.INDICATOR_EPISODE_DIGITS_KEY] = int(self.indicator_episode_digits)
|
||||||
|
|
||||||
|
return ShowDescriptor(**kwargs)
|
||||||
216
src/ffx/model/track.py
Normal file
216
src/ffx/model/track.py
Normal file
@@ -0,0 +1,216 @@
|
|||||||
|
# from typing import List
|
||||||
|
from sqlalchemy import create_engine, Column, Integer, String, ForeignKey
|
||||||
|
from sqlalchemy.orm import relationship, declarative_base, sessionmaker
|
||||||
|
|
||||||
|
from .show import Base
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.iso_language import IsoLanguage
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
|
||||||
|
from ffx.audio_layout import AudioLayout
|
||||||
|
from ffx.track_codec import TrackCodec
|
||||||
|
|
||||||
|
|
||||||
|
class Track(Base):
|
||||||
|
"""
|
||||||
|
relationship(argument, opt1, opt2, ...)
|
||||||
|
argument is string of class or Mapped class of the target entity
|
||||||
|
backref creates a bi-directional corresponding relationship (back_populates preferred)
|
||||||
|
back_populates points to the corresponding relationship (the actual class attribute identifier)
|
||||||
|
|
||||||
|
See: https://docs.sqlalchemy.org/en/(14|20)/orm/basic_relationships.html
|
||||||
|
"""
|
||||||
|
|
||||||
|
__tablename__ = 'tracks'
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
id = Column(Integer, primary_key=True, autoincrement = True)
|
||||||
|
|
||||||
|
# P=pattern_id+sub_index+track_type
|
||||||
|
track_type = Column(Integer) # TrackType
|
||||||
|
|
||||||
|
index = Column(Integer)
|
||||||
|
source_index = Column(Integer)
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
pattern_id = Column(Integer, ForeignKey('patterns.id', ondelete="CASCADE"))
|
||||||
|
pattern = relationship('Pattern', back_populates='tracks')
|
||||||
|
|
||||||
|
track_tags = relationship('TrackTag', back_populates='track', cascade="all, delete", lazy="joined")
|
||||||
|
|
||||||
|
disposition_flags = Column(Integer)
|
||||||
|
|
||||||
|
codec_name = Column(String)
|
||||||
|
audio_layout = Column(Integer)
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, **kwargs):
|
||||||
|
|
||||||
|
trackType = kwargs.pop('track_type', None)
|
||||||
|
if trackType is not None:
|
||||||
|
self.track_type = int(trackType)
|
||||||
|
|
||||||
|
dispositionSet = kwargs.pop(TrackDescriptor.DISPOSITION_SET_KEY, set())
|
||||||
|
self.disposition_flags = int(TrackDisposition.toFlags(dispositionSet))
|
||||||
|
|
||||||
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def fromFfprobeStreamObj(cls, streamObj, patternId):
|
||||||
|
"""{
|
||||||
|
'index': 4,
|
||||||
|
'codec_name': 'hdmv_pgs_subtitle',
|
||||||
|
'codec_long_name': 'HDMV Presentation Graphic Stream subtitles',
|
||||||
|
'codec_type': 'subtitle',
|
||||||
|
'codec_tag_string': '[0][0][0][0]',
|
||||||
|
'codec_tag': '0x0000',
|
||||||
|
'r_frame_rate': '0/0',
|
||||||
|
'avg_frame_rate': '0/0',
|
||||||
|
'time_base': '1/1000',
|
||||||
|
'start_pts': 0,
|
||||||
|
'start_time': '0.000000',
|
||||||
|
'duration_ts': 1421035,
|
||||||
|
'duration': '1421.035000',
|
||||||
|
'disposition': {
|
||||||
|
'default': 1,
|
||||||
|
'dub': 0,
|
||||||
|
'original': 0,
|
||||||
|
'comment': 0,
|
||||||
|
'lyrics': 0,
|
||||||
|
'karaoke': 0,
|
||||||
|
'forced': 0,
|
||||||
|
'hearing_impaired': 0,
|
||||||
|
'visual_impaired': 0,
|
||||||
|
'clean_effects': 0,
|
||||||
|
'attached_pic': 0,
|
||||||
|
'timed_thumbnails': 0,
|
||||||
|
'non_diegetic': 0,
|
||||||
|
'captions': 0,
|
||||||
|
'descriptions': 0,
|
||||||
|
'metadata': 0,
|
||||||
|
'dependent': 0,
|
||||||
|
'still_image': 0
|
||||||
|
},
|
||||||
|
'tags': {
|
||||||
|
'language': 'ger',
|
||||||
|
'title': 'German Full'
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
id = Column(Integer, primary_key=True, autoincrement = True)
|
||||||
|
|
||||||
|
# P=pattern_id+sub_index+track_type
|
||||||
|
track_type = Column(Integer) # TrackType
|
||||||
|
sub_index = Column(Integer)
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
pattern_id = Column(Integer, ForeignKey('patterns.id', ondelete='CASCADE'))
|
||||||
|
pattern = relationship('Pattern', back_populates='tracks')
|
||||||
|
|
||||||
|
|
||||||
|
language = Column(String) # IsoLanguage threeLetter
|
||||||
|
title = Column(String)
|
||||||
|
|
||||||
|
|
||||||
|
track_tags = relationship('TrackTag', back_populates='track', cascade='all, delete')
|
||||||
|
|
||||||
|
|
||||||
|
disposition_flags = Column(Integer)
|
||||||
|
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
|
||||||
|
trackType = streamObj[TrackDescriptor.FFPROBE_CODEC_TYPE_KEY]
|
||||||
|
|
||||||
|
if trackType in [t.label() for t in TrackType]:
|
||||||
|
|
||||||
|
return cls(pattern_id = patternId,
|
||||||
|
track_type = trackType,
|
||||||
|
codec_name = streamObj[TrackDescriptor.FFPROBE_CODEC_NAME_KEY],
|
||||||
|
disposition_flags = sum([2**t.index() for (k,v) in streamObj[TrackDescriptor.FFPROBE_DISPOSITION_KEY].items()
|
||||||
|
if v and (t := TrackDisposition.find(k)) is not None]),
|
||||||
|
audio_layout = AudioLayout.identify(streamObj))
|
||||||
|
|
||||||
|
else:
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def getId(self):
|
||||||
|
return int(self.id)
|
||||||
|
|
||||||
|
def getPatternId(self):
|
||||||
|
return int(self.pattern_id)
|
||||||
|
|
||||||
|
def getType(self):
|
||||||
|
return TrackType.fromIndex(self.track_type)
|
||||||
|
|
||||||
|
def getCodec(self) -> TrackCodec:
|
||||||
|
return TrackCodec.identify(self.codec_name)
|
||||||
|
|
||||||
|
def getIndex(self):
|
||||||
|
return int(self.index) if self.index is not None else -1
|
||||||
|
|
||||||
|
def getSourceIndex(self):
|
||||||
|
return int(self.source_index) if self.source_index is not None else -1
|
||||||
|
|
||||||
|
def getLanguage(self):
|
||||||
|
tags = {t.key:t.value for t in self.track_tags}
|
||||||
|
return IsoLanguage.findThreeLetter(tags['language']) if 'language' in tags.keys() else IsoLanguage.UNDEFINED
|
||||||
|
|
||||||
|
def getTitle(self):
|
||||||
|
tags = {t.key:t.value for t in self.track_tags}
|
||||||
|
return tags['title'] if 'title' in tags.keys() else ''
|
||||||
|
|
||||||
|
def getDispositionSet(self):
|
||||||
|
return TrackDisposition.toSet(self.disposition_flags)
|
||||||
|
|
||||||
|
def getAudioLayout(self):
|
||||||
|
return AudioLayout.fromIndex(self.audio_layout)
|
||||||
|
|
||||||
|
def getTags(self):
|
||||||
|
return {str(t.key):str(t.value) for t in self.track_tags}
|
||||||
|
|
||||||
|
|
||||||
|
def setDisposition(self, disposition : TrackDisposition):
|
||||||
|
self.disposition_flags = self.disposition_flags | int(2**disposition.index())
|
||||||
|
|
||||||
|
def resetDisposition(self, disposition : TrackDisposition):
|
||||||
|
self.disposition_flags = self.disposition_flags & sum([2**d.index() for d in TrackDisposition if d != disposition])
|
||||||
|
|
||||||
|
def getDisposition(self, disposition : TrackDisposition):
|
||||||
|
return bool(self.disposition_flags & 2**disposition.index())
|
||||||
|
|
||||||
|
|
||||||
|
def getDescriptor(self, context = None, subIndex : int = -1) -> TrackDescriptor:
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
|
||||||
|
if not context is None:
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = context
|
||||||
|
|
||||||
|
kwargs[TrackDescriptor.ID_KEY] = self.getId()
|
||||||
|
kwargs[TrackDescriptor.PATTERN_ID_KEY] = self.getPatternId()
|
||||||
|
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = self.getIndex()
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = self.getSourceIndex()
|
||||||
|
|
||||||
|
if subIndex > -1:
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = subIndex
|
||||||
|
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = self.getType()
|
||||||
|
kwargs[TrackDescriptor.CODEC_KEY] = self.getCodec()
|
||||||
|
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = self.getDispositionSet()
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = self.getTags()
|
||||||
|
|
||||||
|
kwargs[TrackDescriptor.AUDIO_LAYOUT_KEY] = self.getAudioLayout()
|
||||||
|
|
||||||
|
return TrackDescriptor(**kwargs)
|
||||||
28
src/ffx/model/track_tag.py
Normal file
28
src/ffx/model/track_tag.py
Normal file
@@ -0,0 +1,28 @@
|
|||||||
|
# from typing import List
|
||||||
|
from sqlalchemy import create_engine, Column, Integer, String, ForeignKey, Enum
|
||||||
|
from sqlalchemy.orm import relationship, declarative_base, sessionmaker
|
||||||
|
|
||||||
|
from .show import Base
|
||||||
|
|
||||||
|
|
||||||
|
class TrackTag(Base):
|
||||||
|
"""
|
||||||
|
relationship(argument, opt1, opt2, ...)
|
||||||
|
argument is string of class or Mapped class of the target entity
|
||||||
|
backref creates a bi-directional corresponding relationship (back_populates preferred)
|
||||||
|
back_populates points to the corresponding relationship (the actual class attribute identifier)
|
||||||
|
|
||||||
|
See: https://docs.sqlalchemy.org/en/(14|20)/orm/basic_relationships.html
|
||||||
|
"""
|
||||||
|
|
||||||
|
__tablename__ = 'track_tags'
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
id = Column(Integer, primary_key=True)
|
||||||
|
|
||||||
|
key = Column(String)
|
||||||
|
value = Column(String)
|
||||||
|
|
||||||
|
# v1.x
|
||||||
|
track_id = Column(Integer, ForeignKey('tracks.id', ondelete="CASCADE"))
|
||||||
|
track = relationship('Track', back_populates='track_tags')
|
||||||
156
src/ffx/pattern_controller.py
Normal file
156
src/ffx/pattern_controller.py
Normal file
@@ -0,0 +1,156 @@
|
|||||||
|
import click, re
|
||||||
|
|
||||||
|
from ffx.model.pattern import Pattern
|
||||||
|
|
||||||
|
|
||||||
|
class PatternController():
|
||||||
|
|
||||||
|
def __init__(self, context):
|
||||||
|
|
||||||
|
self.context = context
|
||||||
|
self.Session = self.context['database']['session'] # convenience
|
||||||
|
|
||||||
|
|
||||||
|
def addPattern(self, patternDescriptor):
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Pattern).filter(Pattern.show_id == int(patternDescriptor['show_id']),
|
||||||
|
Pattern.pattern == str(patternDescriptor['pattern']))
|
||||||
|
|
||||||
|
if not q.count():
|
||||||
|
pattern = Pattern(show_id = int(patternDescriptor['show_id']),
|
||||||
|
pattern = str(patternDescriptor['pattern']))
|
||||||
|
s.add(pattern)
|
||||||
|
s.commit()
|
||||||
|
return pattern.getId()
|
||||||
|
else:
|
||||||
|
return 0
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"PatternController.addPattern(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def updatePattern(self, patternId, patternDescriptor):
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Pattern).filter(Pattern.id == int(patternId))
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
|
||||||
|
pattern = q.first()
|
||||||
|
|
||||||
|
pattern.show_id = int(patternDescriptor['show_id'])
|
||||||
|
pattern.pattern = str(patternDescriptor['pattern'])
|
||||||
|
|
||||||
|
s.commit()
|
||||||
|
return True
|
||||||
|
|
||||||
|
else:
|
||||||
|
return False
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"PatternController.updatePattern(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def findPattern(self, patternDescriptor):
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Pattern).filter(Pattern.show_id == int(patternDescriptor['show_id']), Pattern.pattern == str(patternDescriptor['pattern']))
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
pattern = q.first()
|
||||||
|
return int(pattern.id)
|
||||||
|
else:
|
||||||
|
return None
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"PatternController.findPattern(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def getPattern(self, patternId : int):
|
||||||
|
|
||||||
|
if type(patternId) is not int:
|
||||||
|
raise ValueError(f"PatternController.getPattern(): Argument patternId is required to be of type int")
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Pattern).filter(Pattern.id == int(patternId))
|
||||||
|
|
||||||
|
return q.first() if q.count() else None
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"PatternController.getPattern(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def deletePattern(self, patternId):
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Pattern).filter(Pattern.id == int(patternId))
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
|
||||||
|
#DAFUQ: https://stackoverflow.com/a/19245058
|
||||||
|
# q.delete()
|
||||||
|
pattern = q.first()
|
||||||
|
s.delete(pattern)
|
||||||
|
|
||||||
|
s.commit()
|
||||||
|
return True
|
||||||
|
return False
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"PatternController.deletePattern(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def matchFilename(self, filename : str) -> dict:
|
||||||
|
"""Returns dict {'match': <a regex match obj>, 'pattern': <ffx pattern obj>} or empty dict of no pattern was found"""
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Pattern)
|
||||||
|
|
||||||
|
matchResult = {}
|
||||||
|
|
||||||
|
for pattern in q.all():
|
||||||
|
patternMatch = re.search(str(pattern.pattern), str(filename))
|
||||||
|
if patternMatch is not None:
|
||||||
|
matchResult['match'] = patternMatch
|
||||||
|
matchResult['pattern'] = pattern
|
||||||
|
|
||||||
|
return matchResult
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"PatternController.matchFilename(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
# def getMediaDescriptor(self, context, patternId):
|
||||||
|
#
|
||||||
|
# try:
|
||||||
|
# s = self.Session()
|
||||||
|
# q = s.query(Pattern).filter(Pattern.id == int(patternId))
|
||||||
|
#
|
||||||
|
# if q.count():
|
||||||
|
# return q.first().getMediaDescriptor(context)
|
||||||
|
# else:
|
||||||
|
# return None
|
||||||
|
#
|
||||||
|
# except Exception as ex:
|
||||||
|
# raise click.ClickException(f"PatternController.getMediaDescriptor(): {repr(ex)}")
|
||||||
|
# finally:
|
||||||
|
# s.close()
|
||||||
111
src/ffx/pattern_delete_screen.py
Normal file
111
src/ffx/pattern_delete_screen.py
Normal file
@@ -0,0 +1,111 @@
|
|||||||
|
import click
|
||||||
|
|
||||||
|
from textual.screen import Screen
|
||||||
|
from textual.widgets import Header, Footer, Static, Button
|
||||||
|
from textual.containers import Grid
|
||||||
|
|
||||||
|
from .show_controller import ShowController
|
||||||
|
from .pattern_controller import PatternController
|
||||||
|
|
||||||
|
from ffx.model.pattern import Pattern
|
||||||
|
|
||||||
|
|
||||||
|
# Screen[dict[int, str, int]]
|
||||||
|
class PatternDeleteScreen(Screen):
|
||||||
|
|
||||||
|
CSS = """
|
||||||
|
|
||||||
|
Grid {
|
||||||
|
grid-size: 2;
|
||||||
|
grid-rows: 2 auto;
|
||||||
|
grid-columns: 30 330;
|
||||||
|
height: 100%;
|
||||||
|
width: 100%;
|
||||||
|
padding: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
Input {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
Button {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
#toplabel {
|
||||||
|
height: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
.two {
|
||||||
|
column-span: 2;
|
||||||
|
}
|
||||||
|
|
||||||
|
.box {
|
||||||
|
height: 100%;
|
||||||
|
border: solid green;
|
||||||
|
}
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, patternId = None, showId = None):
|
||||||
|
super().__init__()
|
||||||
|
|
||||||
|
self.context = self.app.getContext()
|
||||||
|
self.Session = self.context['database']['session'] # convenience
|
||||||
|
|
||||||
|
self.__pc = PatternController(context = self.context)
|
||||||
|
self.__sc = ShowController(context = self.context)
|
||||||
|
|
||||||
|
self.__patternId = patternId
|
||||||
|
self.__pattern: Pattern = self.__pc.getPattern(patternId) if patternId is not None else {}
|
||||||
|
self.__showDescriptor = self.__sc.getShowDescriptor(showId) if showId is not None else {}
|
||||||
|
|
||||||
|
|
||||||
|
def on_mount(self):
|
||||||
|
if self.__showDescriptor:
|
||||||
|
self.query_one("#showlabel", Static).update(f"{self.__showDescriptor.getId()} - {self.__showDescriptor.getName()} ({self.__showDescriptor.getYear()})")
|
||||||
|
if not self.__pattern is None:
|
||||||
|
self.query_one("#patternlabel", Static).update(str(self.__pattern.pattern))
|
||||||
|
|
||||||
|
|
||||||
|
def compose(self):
|
||||||
|
|
||||||
|
yield Header()
|
||||||
|
|
||||||
|
with Grid():
|
||||||
|
|
||||||
|
yield Static("Are you sure to delete the following filename pattern?", id="toplabel", classes="two")
|
||||||
|
|
||||||
|
yield Static("", classes="two")
|
||||||
|
|
||||||
|
yield Static("Pattern")
|
||||||
|
yield Static("", id="patternlabel")
|
||||||
|
|
||||||
|
yield Static("", classes="two")
|
||||||
|
|
||||||
|
yield Static("from show")
|
||||||
|
yield Static("", id="showlabel")
|
||||||
|
|
||||||
|
yield Static("", classes="two")
|
||||||
|
|
||||||
|
yield Button("Delete", id="delete_button")
|
||||||
|
yield Button("Cancel", id="cancel_button")
|
||||||
|
|
||||||
|
yield Footer()
|
||||||
|
|
||||||
|
|
||||||
|
# Event handler for button press
|
||||||
|
def on_button_pressed(self, event: Button.Pressed) -> None:
|
||||||
|
|
||||||
|
if event.button.id == "delete_button":
|
||||||
|
|
||||||
|
if self.__patternId is None:
|
||||||
|
raise click.ClickException('PatternDeleteScreen.on_button_pressed(): pattern id is undefined')
|
||||||
|
|
||||||
|
if self.__pc.deletePattern(self.__patternId):
|
||||||
|
self.dismiss(self.__pattern)
|
||||||
|
|
||||||
|
else:
|
||||||
|
#TODO: Meldung
|
||||||
|
self.app.pop_screen()
|
||||||
|
|
||||||
|
if event.button.id == "cancel_button":
|
||||||
|
self.app.pop_screen()
|
||||||
|
|
||||||
548
src/ffx/pattern_details_screen.py
Normal file
548
src/ffx/pattern_details_screen.py
Normal file
@@ -0,0 +1,548 @@
|
|||||||
|
import click, re
|
||||||
|
from typing import List
|
||||||
|
|
||||||
|
from textual.screen import Screen
|
||||||
|
from textual.widgets import Header, Footer, Static, Button, Input, DataTable
|
||||||
|
from textual.containers import Grid
|
||||||
|
|
||||||
|
from ffx.model.pattern import Pattern
|
||||||
|
from ffx.model.track import Track
|
||||||
|
|
||||||
|
from .pattern_controller import PatternController
|
||||||
|
from .show_controller import ShowController
|
||||||
|
from .track_controller import TrackController
|
||||||
|
from .tag_controller import TagController
|
||||||
|
|
||||||
|
from .track_details_screen import TrackDetailsScreen
|
||||||
|
from .track_delete_screen import TrackDeleteScreen
|
||||||
|
|
||||||
|
from .tag_details_screen import TagDetailsScreen
|
||||||
|
from .tag_delete_screen import TagDeleteScreen
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
|
||||||
|
from textual.widgets._data_table import CellDoesNotExist
|
||||||
|
|
||||||
|
from ffx.file_properties import FileProperties
|
||||||
|
from ffx.iso_language import IsoLanguage
|
||||||
|
from ffx.audio_layout import AudioLayout
|
||||||
|
|
||||||
|
|
||||||
|
# Screen[dict[int, str, int]]
|
||||||
|
class PatternDetailsScreen(Screen):
|
||||||
|
|
||||||
|
CSS = """
|
||||||
|
|
||||||
|
Grid {
|
||||||
|
grid-size: 7 13;
|
||||||
|
grid-rows: 2 2 2 2 2 8 2 2 8 2 2 2 2;
|
||||||
|
grid-columns: 25 25 25 25 25 25 25;
|
||||||
|
height: 100%;
|
||||||
|
width: 100%;
|
||||||
|
padding: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
Input {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
Button {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable {
|
||||||
|
min-height: 6;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable .datatable--cursor {
|
||||||
|
background: darkorange;
|
||||||
|
color: black;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable .datatable--header {
|
||||||
|
background: steelblue;
|
||||||
|
color: white;
|
||||||
|
}
|
||||||
|
|
||||||
|
#toplabel {
|
||||||
|
height: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
.three {
|
||||||
|
column-span: 3;
|
||||||
|
}
|
||||||
|
|
||||||
|
.four {
|
||||||
|
column-span: 4;
|
||||||
|
}
|
||||||
|
.five {
|
||||||
|
column-span: 5;
|
||||||
|
}
|
||||||
|
.six {
|
||||||
|
column-span: 6;
|
||||||
|
}
|
||||||
|
.seven {
|
||||||
|
column-span: 7;
|
||||||
|
}
|
||||||
|
|
||||||
|
.box {
|
||||||
|
height: 100%;
|
||||||
|
border: solid green;
|
||||||
|
}
|
||||||
|
|
||||||
|
.yellow {
|
||||||
|
tint: yellow 40%;
|
||||||
|
}
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, patternId = None, showId = None):
|
||||||
|
super().__init__()
|
||||||
|
|
||||||
|
self.context = self.app.getContext()
|
||||||
|
self.Session = self.context['database']['session'] # convenience
|
||||||
|
|
||||||
|
self.__pc = PatternController(context = self.context)
|
||||||
|
self.__sc = ShowController(context = self.context)
|
||||||
|
self.__tc = TrackController(context = self.context)
|
||||||
|
self.__tac = TagController(context = self.context)
|
||||||
|
|
||||||
|
self.__pattern : Pattern = self.__pc.getPattern(patternId) if patternId is not None else None
|
||||||
|
self.__showDescriptor = self.__sc.getShowDescriptor(showId) if showId is not None else None
|
||||||
|
|
||||||
|
|
||||||
|
#TODO: per controller
|
||||||
|
def loadTracks(self, show_id):
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
tracks = {}
|
||||||
|
tracks['audio'] = {}
|
||||||
|
tracks['subtitle'] = {}
|
||||||
|
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Pattern).filter(Pattern.show_id == int(show_id))
|
||||||
|
|
||||||
|
return [{'id': int(p.id), 'pattern': p.pattern} for p in q.all()]
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"loadTracks(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def updateTracks(self):
|
||||||
|
|
||||||
|
self.tracksTable.clear()
|
||||||
|
|
||||||
|
if self.__pattern is not None:
|
||||||
|
|
||||||
|
tracks = self.__tc.findTracks(self.__pattern.getId())
|
||||||
|
|
||||||
|
typeCounter = {}
|
||||||
|
|
||||||
|
tr: Track
|
||||||
|
for tr in tracks:
|
||||||
|
|
||||||
|
td : TrackDescriptor = tr.getDescriptor(self.context)
|
||||||
|
|
||||||
|
trackType = td.getType()
|
||||||
|
if not trackType in typeCounter.keys():
|
||||||
|
typeCounter[trackType] = 0
|
||||||
|
|
||||||
|
dispoSet = td.getDispositionSet()
|
||||||
|
|
||||||
|
trackLanguage = td.getLanguage()
|
||||||
|
audioLayout = td.getAudioLayout()
|
||||||
|
row = (td.getIndex(),
|
||||||
|
trackType.label(),
|
||||||
|
typeCounter[trackType],
|
||||||
|
td.getCodec().label(),
|
||||||
|
audioLayout.label() if trackType == TrackType.AUDIO
|
||||||
|
and audioLayout != AudioLayout.LAYOUT_UNDEFINED else ' ',
|
||||||
|
trackLanguage.label() if trackLanguage != IsoLanguage.UNDEFINED else ' ',
|
||||||
|
td.getTitle(),
|
||||||
|
'Yes' if TrackDisposition.DEFAULT in dispoSet else 'No',
|
||||||
|
'Yes' if TrackDisposition.FORCED in dispoSet else 'No',
|
||||||
|
td.getSourceIndex())
|
||||||
|
|
||||||
|
self.tracksTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
typeCounter[trackType] += 1
|
||||||
|
|
||||||
|
|
||||||
|
def swapTracks(self, trackIndex1: int, trackIndex2: int):
|
||||||
|
|
||||||
|
ti1 = int(trackIndex1)
|
||||||
|
ti2 = int(trackIndex2)
|
||||||
|
|
||||||
|
siblingDescriptors: List[TrackDescriptor] = self.__tc.findSiblingDescriptors(self.__pattern.getId())
|
||||||
|
|
||||||
|
numSiblings = len(siblingDescriptors)
|
||||||
|
|
||||||
|
if ti1 < 0 or ti1 >= numSiblings:
|
||||||
|
raise ValueError(f"PatternDetailsScreen.swapTracks(): trackIndex1 ({ti1}) is out of range ({numSiblings})")
|
||||||
|
|
||||||
|
if ti2 < 0 or ti2 >= numSiblings:
|
||||||
|
raise ValueError(f"PatternDetailsScreen.swapTracks(): trackIndex2 ({ti2}) is out of range ({numSiblings})")
|
||||||
|
|
||||||
|
sibling1 = siblingDescriptors[trackIndex1]
|
||||||
|
sibling2 = siblingDescriptors[trackIndex2]
|
||||||
|
|
||||||
|
# raise click.ClickException(f"siblings id1={sibling1.getId()} id2={sibling2.getId()}")
|
||||||
|
|
||||||
|
subIndex2 = sibling2.getSubIndex()
|
||||||
|
|
||||||
|
sibling2.setIndex(sibling1.getIndex())
|
||||||
|
sibling2.setSubIndex(sibling1.getSubIndex())
|
||||||
|
|
||||||
|
sibling1.setIndex(trackIndex2)
|
||||||
|
sibling1.setSubIndex(subIndex2)
|
||||||
|
|
||||||
|
if not self.__tc.updateTrack(sibling1.getId(), sibling1):
|
||||||
|
raise click.ClickException('Update sibling1 failed')
|
||||||
|
if not self.__tc.updateTrack(sibling2.getId(), sibling2):
|
||||||
|
raise click.ClickException('Update sibling2 failed')
|
||||||
|
|
||||||
|
self.updateTracks()
|
||||||
|
|
||||||
|
|
||||||
|
def updateTags(self):
|
||||||
|
|
||||||
|
self.tagsTable.clear()
|
||||||
|
|
||||||
|
if self.__pattern is not None:
|
||||||
|
|
||||||
|
tags = self.__tac.findAllMediaTags(self.__pattern.getId())
|
||||||
|
|
||||||
|
for tagKey, tagValue in tags.items():
|
||||||
|
row = (tagKey, tagValue)
|
||||||
|
self.tagsTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
|
||||||
|
def on_mount(self):
|
||||||
|
|
||||||
|
if not self.__showDescriptor is None:
|
||||||
|
self.query_one("#showlabel", Static).update(f"{self.__showDescriptor.getId()} - {self.__showDescriptor.getName()} ({self.__showDescriptor.getYear()})")
|
||||||
|
|
||||||
|
if self.__pattern is not None:
|
||||||
|
|
||||||
|
self.query_one("#pattern_input", Input).value = str(self.__pattern.getPattern())
|
||||||
|
|
||||||
|
self.updateTags()
|
||||||
|
self.updateTracks()
|
||||||
|
|
||||||
|
def compose(self):
|
||||||
|
|
||||||
|
|
||||||
|
self.tagsTable = DataTable(classes="seven")
|
||||||
|
|
||||||
|
# Define the columns with headers
|
||||||
|
self.column_key_tag_key = self.tagsTable.add_column("Key", width=50)
|
||||||
|
self.column_key_tag_value = self.tagsTable.add_column("Value", width=100)
|
||||||
|
|
||||||
|
self.tagsTable.cursor_type = 'row'
|
||||||
|
|
||||||
|
|
||||||
|
self.tracksTable = DataTable(id="tracks_table", classes="seven")
|
||||||
|
|
||||||
|
self.column_key_track_index = self.tracksTable.add_column("Index", width=5)
|
||||||
|
self.column_key_track_type = self.tracksTable.add_column("Type", width=10)
|
||||||
|
self.column_key_track_sub_index = self.tracksTable.add_column("SubIndex", width=8)
|
||||||
|
self.column_key_track_codec = self.tracksTable.add_column("Codec", width=10)
|
||||||
|
self.column_key_track_audio_layout = self.tracksTable.add_column("Layout", width=10)
|
||||||
|
self.column_key_track_language = self.tracksTable.add_column("Language", width=15)
|
||||||
|
self.column_key_track_title = self.tracksTable.add_column("Title", width=48)
|
||||||
|
self.column_key_track_default = self.tracksTable.add_column("Default", width=8)
|
||||||
|
self.column_key_track_forced = self.tracksTable.add_column("Forced", width=8)
|
||||||
|
self.column_key_track_source_index = self.tracksTable.add_column("SrcIndex", width=8)
|
||||||
|
|
||||||
|
self.tracksTable.cursor_type = 'row'
|
||||||
|
|
||||||
|
|
||||||
|
yield Header()
|
||||||
|
|
||||||
|
with Grid():
|
||||||
|
|
||||||
|
# 1
|
||||||
|
yield Static("Edit filename pattern" if self.__pattern is not None else "New filename pattern", id="toplabel")
|
||||||
|
yield Input(type="text", id="pattern_input", classes="six")
|
||||||
|
|
||||||
|
# 2
|
||||||
|
yield Static("from show")
|
||||||
|
yield Static("", id="showlabel", classes="five")
|
||||||
|
yield Button("Substitute pattern", id="pattern_button")
|
||||||
|
|
||||||
|
# 3
|
||||||
|
yield Static(" ", classes="seven")
|
||||||
|
# 4
|
||||||
|
yield Static(" ", classes="seven")
|
||||||
|
|
||||||
|
# 5
|
||||||
|
yield Static("Media Tags")
|
||||||
|
|
||||||
|
|
||||||
|
if self.__pattern is not None:
|
||||||
|
yield Button("Add", id="button_add_tag")
|
||||||
|
yield Button("Edit", id="button_edit_tag")
|
||||||
|
yield Button("Delete", id="button_delete_tag")
|
||||||
|
else:
|
||||||
|
yield Static(" ")
|
||||||
|
yield Static(" ")
|
||||||
|
yield Static(" ")
|
||||||
|
|
||||||
|
yield Static(" ")
|
||||||
|
yield Static(" ")
|
||||||
|
yield Static(" ")
|
||||||
|
|
||||||
|
# 6
|
||||||
|
yield self.tagsTable
|
||||||
|
|
||||||
|
# 7
|
||||||
|
yield Static(" ", classes="seven")
|
||||||
|
|
||||||
|
# 8
|
||||||
|
yield Static("Streams")
|
||||||
|
|
||||||
|
|
||||||
|
if self.__pattern is not None:
|
||||||
|
yield Button("Add", id="button_add_track")
|
||||||
|
yield Button("Edit", id="button_edit_track")
|
||||||
|
yield Button("Delete", id="button_delete_track")
|
||||||
|
else:
|
||||||
|
yield Static(" ")
|
||||||
|
yield Static(" ")
|
||||||
|
yield Static(" ")
|
||||||
|
|
||||||
|
yield Static(" ")
|
||||||
|
yield Button("Up", id="button_track_up")
|
||||||
|
yield Button("Down", id="button_track_down")
|
||||||
|
|
||||||
|
# 9
|
||||||
|
yield self.tracksTable
|
||||||
|
|
||||||
|
# 10
|
||||||
|
yield Static(" ", classes="seven")
|
||||||
|
|
||||||
|
# 11
|
||||||
|
yield Static(" ", classes="seven")
|
||||||
|
|
||||||
|
# 12
|
||||||
|
yield Button("Save", id="save_button")
|
||||||
|
yield Button("Cancel", id="cancel_button")
|
||||||
|
yield Static(" ", classes="five")
|
||||||
|
|
||||||
|
# 13
|
||||||
|
yield Static(" ", classes="seven")
|
||||||
|
|
||||||
|
yield Footer()
|
||||||
|
|
||||||
|
|
||||||
|
def getPatternFromInput(self):
|
||||||
|
return str(self.query_one("#pattern_input", Input).value)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def getSelectedTrackDescriptor(self):
|
||||||
|
|
||||||
|
if not self.__pattern:
|
||||||
|
return None
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
# Fetch the currently selected row when 'Enter' is pressed
|
||||||
|
#selected_row_index = self.table.cursor_row
|
||||||
|
row_key, col_key = self.tracksTable.coordinate_to_cell_key(self.tracksTable.cursor_coordinate)
|
||||||
|
|
||||||
|
if row_key is not None:
|
||||||
|
selected_track_data = self.tracksTable.get_row(row_key)
|
||||||
|
|
||||||
|
trackIndex = int(selected_track_data[0])
|
||||||
|
trackSubIndex = int(selected_track_data[2])
|
||||||
|
|
||||||
|
return self.__tc.getTrack(self.__pattern.getId(), trackIndex).getDescriptor(self.context, subIndex=trackSubIndex)
|
||||||
|
|
||||||
|
else:
|
||||||
|
return None
|
||||||
|
|
||||||
|
except CellDoesNotExist:
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def getSelectedTag(self):
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
# Fetch the currently selected row when 'Enter' is pressed
|
||||||
|
#selected_row_index = self.table.cursor_row
|
||||||
|
row_key, col_key = self.tagsTable.coordinate_to_cell_key(self.tagsTable.cursor_coordinate)
|
||||||
|
|
||||||
|
if row_key is not None:
|
||||||
|
selected_tag_data = self.tagsTable.get_row(row_key)
|
||||||
|
|
||||||
|
tagKey = str(selected_tag_data[0])
|
||||||
|
tagValue = str(selected_tag_data[1])
|
||||||
|
|
||||||
|
return tagKey, tagValue
|
||||||
|
|
||||||
|
else:
|
||||||
|
return None
|
||||||
|
|
||||||
|
except CellDoesNotExist:
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
# Event handler for button press
|
||||||
|
def on_button_pressed(self, event: Button.Pressed) -> None:
|
||||||
|
# Check if the button pressed is the one we are interested in
|
||||||
|
if event.button.id == "save_button":
|
||||||
|
|
||||||
|
patternDescriptor = {}
|
||||||
|
patternDescriptor['show_id'] = self.__showDescriptor.getId()
|
||||||
|
patternDescriptor['pattern'] = self.getPatternFromInput()
|
||||||
|
|
||||||
|
if self.__pattern is not None:
|
||||||
|
|
||||||
|
if self.__pc.updatePattern(self.__pattern.getId(), patternDescriptor):
|
||||||
|
self.dismiss(patternDescriptor)
|
||||||
|
else:
|
||||||
|
#TODO: Meldung
|
||||||
|
self.app.pop_screen()
|
||||||
|
|
||||||
|
else:
|
||||||
|
patternId = self.__pc.addPattern(patternDescriptor)
|
||||||
|
if patternId:
|
||||||
|
self.dismiss(patternDescriptor)
|
||||||
|
else:
|
||||||
|
#TODO: Meldung
|
||||||
|
self.app.pop_screen()
|
||||||
|
|
||||||
|
|
||||||
|
if event.button.id == "cancel_button":
|
||||||
|
self.app.pop_screen()
|
||||||
|
|
||||||
|
|
||||||
|
# Save pattern when just created before adding streams
|
||||||
|
if self.__pattern is not None:
|
||||||
|
|
||||||
|
numTracks = len(self.tracksTable.rows)
|
||||||
|
|
||||||
|
if event.button.id == "button_add_track":
|
||||||
|
self.app.push_screen(TrackDetailsScreen(patternId = self.__pattern.getId(), index = numTracks), self.handle_add_track)
|
||||||
|
|
||||||
|
selectedTrack = self.getSelectedTrackDescriptor()
|
||||||
|
if selectedTrack is not None:
|
||||||
|
if event.button.id == "button_edit_track":
|
||||||
|
self.app.push_screen(TrackDetailsScreen(trackDescriptor = selectedTrack), self.handle_edit_track)
|
||||||
|
if event.button.id == "button_delete_track":
|
||||||
|
self.app.push_screen(TrackDeleteScreen(trackDescriptor = selectedTrack), self.handle_delete_track)
|
||||||
|
|
||||||
|
|
||||||
|
if event.button.id == "button_add_tag":
|
||||||
|
if self.__pattern is not None:
|
||||||
|
self.app.push_screen(TagDetailsScreen(), self.handle_update_tag)
|
||||||
|
|
||||||
|
if event.button.id == "button_edit_tag":
|
||||||
|
tagKey, tagValue = self.getSelectedTag()
|
||||||
|
self.app.push_screen(TagDetailsScreen(key=tagKey, value=tagValue), self.handle_update_tag)
|
||||||
|
|
||||||
|
if event.button.id == "button_delete_tag":
|
||||||
|
tagKey, tagValue = self.getSelectedTag()
|
||||||
|
self.app.push_screen(TagDeleteScreen(key=tagKey, value=tagValue), self.handle_delete_tag)
|
||||||
|
|
||||||
|
|
||||||
|
if event.button.id == "pattern_button":
|
||||||
|
|
||||||
|
pattern = self.query_one("#pattern_input", Input).value
|
||||||
|
|
||||||
|
patternMatch = re.search(FileProperties.SE_INDICATOR_PATTERN, pattern)
|
||||||
|
|
||||||
|
if patternMatch:
|
||||||
|
self.query_one("#pattern_input", Input).value = pattern.replace(patternMatch.group(1),
|
||||||
|
FileProperties.SE_INDICATOR_PATTERN)
|
||||||
|
|
||||||
|
|
||||||
|
if event.button.id == "button_track_up":
|
||||||
|
|
||||||
|
selectedTrackDescriptor = self.getSelectedTrackDescriptor()
|
||||||
|
selectedTrackIndex = selectedTrackDescriptor.getIndex()
|
||||||
|
|
||||||
|
if selectedTrackIndex > 0 and selectedTrackIndex < self.tracksTable.row_count:
|
||||||
|
correspondingTrackIndex = selectedTrackIndex - 1
|
||||||
|
self.swapTracks(selectedTrackIndex, correspondingTrackIndex)
|
||||||
|
|
||||||
|
|
||||||
|
if event.button.id == "button_track_down":
|
||||||
|
|
||||||
|
selectedTrackDescriptor = self.getSelectedTrackDescriptor()
|
||||||
|
selectedTrackIndex = selectedTrackDescriptor.getIndex()
|
||||||
|
|
||||||
|
if selectedTrackIndex >= 0 and selectedTrackIndex < (self.tracksTable.row_count - 1):
|
||||||
|
correspondingTrackIndex = selectedTrackIndex + 1
|
||||||
|
self.swapTracks(selectedTrackIndex, correspondingTrackIndex)
|
||||||
|
|
||||||
|
|
||||||
|
def handle_add_track(self, trackDescriptor : TrackDescriptor):
|
||||||
|
|
||||||
|
dispoSet = trackDescriptor.getDispositionSet()
|
||||||
|
trackType = trackDescriptor.getType()
|
||||||
|
index = trackDescriptor.getIndex()
|
||||||
|
subIndex = trackDescriptor.getSubIndex()
|
||||||
|
codec = trackDescriptor.getCodec()
|
||||||
|
language = trackDescriptor.getLanguage()
|
||||||
|
title = trackDescriptor.getTitle()
|
||||||
|
|
||||||
|
row = (index,
|
||||||
|
trackType.label(),
|
||||||
|
subIndex,
|
||||||
|
codec.label(),
|
||||||
|
language.label(),
|
||||||
|
title,
|
||||||
|
'Yes' if TrackDisposition.DEFAULT in dispoSet else 'No',
|
||||||
|
'Yes' if TrackDisposition.FORCED in dispoSet else 'No')
|
||||||
|
|
||||||
|
self.tracksTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
|
||||||
|
def handle_edit_track(self, trackDescriptor : TrackDescriptor):
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
row_key, col_key = self.tracksTable.coordinate_to_cell_key(self.tracksTable.cursor_coordinate)
|
||||||
|
|
||||||
|
self.tracksTable.update_cell(row_key, self.column_key_track_audio_layout,
|
||||||
|
trackDescriptor.getAudioLayout().label()
|
||||||
|
if trackDescriptor.getType() == TrackType.AUDIO else ' ')
|
||||||
|
|
||||||
|
self.tracksTable.update_cell(row_key, self.column_key_track_language, trackDescriptor.getLanguage().label())
|
||||||
|
self.tracksTable.update_cell(row_key, self.column_key_track_title, trackDescriptor.getTitle())
|
||||||
|
self.tracksTable.update_cell(row_key, self.column_key_track_default, 'Yes' if TrackDisposition.DEFAULT in trackDescriptor.getDispositionSet() else 'No')
|
||||||
|
self.tracksTable.update_cell(row_key, self.column_key_track_forced, 'Yes' if TrackDisposition.FORCED in trackDescriptor.getDispositionSet() else 'No')
|
||||||
|
|
||||||
|
except CellDoesNotExist:
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
def handle_delete_track(self, trackDescriptor : TrackDescriptor):
|
||||||
|
self.updateTracks()
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def handle_update_tag(self, tag):
|
||||||
|
|
||||||
|
if self.__pattern is None:
|
||||||
|
raise click.ClickException(f"PatternDetailsScreen.handle_update_tag: pattern not set")
|
||||||
|
|
||||||
|
if self.__tac.updateMediaTag(self.__pattern.getId(), tag[0], tag[1]) is not None:
|
||||||
|
self.updateTags()
|
||||||
|
|
||||||
|
def handle_delete_tag(self, tag):
|
||||||
|
|
||||||
|
if self.__pattern is None:
|
||||||
|
raise click.ClickException(f"PatternDetailsScreen.handle_delete_tag: pattern not set")
|
||||||
|
|
||||||
|
if self.__tac.deleteMediaTagByKey(self.__pattern.getId(), tag[0]):
|
||||||
|
self.updateTags()
|
||||||
39
src/ffx/process.py
Normal file
39
src/ffx/process.py
Normal file
@@ -0,0 +1,39 @@
|
|||||||
|
import subprocess, logging
|
||||||
|
from typing import List
|
||||||
|
|
||||||
|
def executeProcess(commandSequence: List[str], directory: str = None, context: dict = None):
|
||||||
|
"""
|
||||||
|
niceness -20 bis +19
|
||||||
|
cpu_percent: 1 bis 99
|
||||||
|
"""
|
||||||
|
|
||||||
|
if context is None:
|
||||||
|
logger = logging.getLogger('FFX')
|
||||||
|
logger.addHandler(logging.NullHandler())
|
||||||
|
else:
|
||||||
|
logger = context['logger']
|
||||||
|
|
||||||
|
niceSequence = []
|
||||||
|
|
||||||
|
niceness = (int(context['resource_limits']['niceness'])
|
||||||
|
if not context is None
|
||||||
|
and 'resource_limits' in context.keys()
|
||||||
|
and 'niceness' in context['resource_limits'].keys() else 99)
|
||||||
|
cpu_percent = (int(context['resource_limits']['cpu_percent'])
|
||||||
|
if not context is None
|
||||||
|
and 'resource_limits' in context.keys()
|
||||||
|
and 'cpu_percent' in context['resource_limits'].keys() else 0)
|
||||||
|
|
||||||
|
if niceness >= -20 and niceness <= 19:
|
||||||
|
niceSequence += ['nice', '-n', str(niceness)]
|
||||||
|
if cpu_percent >= 1:
|
||||||
|
niceSequence += ['cpulimit', '-l', str(cpu_percent), '--']
|
||||||
|
|
||||||
|
niceCommand = niceSequence + commandSequence
|
||||||
|
|
||||||
|
logger.debug(f"executeProcess() command sequence: {' '.join(niceCommand)}")
|
||||||
|
|
||||||
|
process = subprocess.Popen(niceCommand, stdout=subprocess.PIPE, stderr=subprocess.PIPE, encoding='utf-8', cwd = directory)
|
||||||
|
output, error = process.communicate()
|
||||||
|
|
||||||
|
return output, error, process.returncode
|
||||||
12
src/ffx/settings_screen.py
Normal file
12
src/ffx/settings_screen.py
Normal file
@@ -0,0 +1,12 @@
|
|||||||
|
from textual.app import ComposeResult
|
||||||
|
from textual.screen import Screen
|
||||||
|
from textual.widgets import Footer, Placeholder
|
||||||
|
|
||||||
|
|
||||||
|
class SettingsScreen(Screen):
|
||||||
|
def __init__(self):
|
||||||
|
super().__init__()
|
||||||
|
context = self.app.getContext()
|
||||||
|
def compose(self) -> ComposeResult:
|
||||||
|
yield Placeholder("Settings Screen")
|
||||||
|
yield Footer()
|
||||||
223
src/ffx/shifted_season_controller.py
Normal file
223
src/ffx/shifted_season_controller.py
Normal file
@@ -0,0 +1,223 @@
|
|||||||
|
import click
|
||||||
|
|
||||||
|
from ffx.model.shifted_season import ShiftedSeason
|
||||||
|
|
||||||
|
|
||||||
|
class EpisodeOrderException(Exception):
|
||||||
|
pass
|
||||||
|
|
||||||
|
class RangeOverlapException(Exception):
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
class ShiftedSeasonController():
|
||||||
|
|
||||||
|
def __init__(self, context):
|
||||||
|
|
||||||
|
self.context = context
|
||||||
|
self.Session = self.context['database']['session'] # convenience
|
||||||
|
|
||||||
|
def checkShiftedSeason(self, showId: int, shiftedSeasonObj: dict, shiftedSeasonId: int = 0):
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
|
||||||
|
firstEpisode = int(shiftedSeasonObj['first_episode'])
|
||||||
|
lastEpisode = int(shiftedSeasonObj['last_episode'])
|
||||||
|
|
||||||
|
q = s.query(ShiftedSeason).filter(ShiftedSeason.show_id == int(showId))
|
||||||
|
if shiftedSeasonId:
|
||||||
|
q = q.filter(ShiftedSeason.id != int(shiftedSeasonId))
|
||||||
|
|
||||||
|
siblingShiftedSeason: ShiftedSeason
|
||||||
|
for siblingShiftedSeason in q.all():
|
||||||
|
|
||||||
|
siblingFirstEpisode = siblingShiftedSeason.getFirstEpisode()
|
||||||
|
siblingLastEpisode = siblingShiftedSeason.getLastEpisode()
|
||||||
|
|
||||||
|
if (lastEpisode >= siblingFirstEpisode
|
||||||
|
and siblingLastEpisode >= firstEpisode):
|
||||||
|
return False
|
||||||
|
return True
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"ShiftedSeasonController.addShiftedSeason(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def addShiftedSeason(self, showId: int, shiftedSeasonObj: dict):
|
||||||
|
|
||||||
|
if type(showId) is not int:
|
||||||
|
raise ValueError(f"ShiftedSeasonController.addShiftedSeason(): Argument showId is required to be of type int")
|
||||||
|
|
||||||
|
if type(shiftedSeasonObj) is not dict:
|
||||||
|
raise ValueError(f"ShiftedSeasonController.addShiftedSeason(): Argument shiftedSeasonObj is required to be of type dict")
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
|
||||||
|
firstEpisode = int(shiftedSeasonObj['first_episode'])
|
||||||
|
lastEpisode = int(shiftedSeasonObj['last_episode'])
|
||||||
|
|
||||||
|
if lastEpisode < firstEpisode:
|
||||||
|
raise EpisodeOrderException()
|
||||||
|
|
||||||
|
q = s.query(ShiftedSeason).filter(ShiftedSeason.show_id == int(showId))
|
||||||
|
|
||||||
|
shiftedSeason = ShiftedSeason(show_id = int(showId),
|
||||||
|
original_season = int(shiftedSeasonObj['original_season']),
|
||||||
|
first_episode = firstEpisode,
|
||||||
|
last_episode = lastEpisode,
|
||||||
|
season_offset = int(shiftedSeasonObj['season_offset']),
|
||||||
|
episode_offset = int(shiftedSeasonObj['episode_offset']))
|
||||||
|
s.add(shiftedSeason)
|
||||||
|
s.commit()
|
||||||
|
return shiftedSeason.getId()
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"ShiftedSeasonController.addShiftedSeason(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def updateShiftedSeason(self, shiftedSeasonId: int, shiftedSeasonObj: dict):
|
||||||
|
|
||||||
|
if type(shiftedSeasonId) is not int:
|
||||||
|
raise ValueError(f"ShiftedSeasonController.updateShiftedSeason(): Argument shiftedSeasonId is required to be of type int")
|
||||||
|
|
||||||
|
if type(shiftedSeasonObj) is not dict:
|
||||||
|
raise ValueError(f"ShiftedSeasonController.updateShiftedSeason(): Argument shiftedSeasonObj is required to be of type dict")
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
|
||||||
|
q = s.query(ShiftedSeason).filter(ShiftedSeason.id == int(shiftedSeasonId))
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
|
||||||
|
shiftedSeason = q.first()
|
||||||
|
|
||||||
|
shiftedSeason.original_season = int(shiftedSeasonObj['original_season'])
|
||||||
|
shiftedSeason.first_episode = int(shiftedSeasonObj['first_episode'])
|
||||||
|
shiftedSeason.last_episode = int(shiftedSeasonObj['last_episode'])
|
||||||
|
shiftedSeason.season_offset = int(shiftedSeasonObj['season_offset'])
|
||||||
|
shiftedSeason.episode_offset = int(shiftedSeasonObj['episode_offset'])
|
||||||
|
|
||||||
|
s.commit()
|
||||||
|
return True
|
||||||
|
|
||||||
|
else:
|
||||||
|
return False
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"ShiftedSeasonController.updateShiftedSeason(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def findShiftedSeason(self, showId: int, originalSeason: int, firstEpisode: int, lastEpisode: int):
|
||||||
|
|
||||||
|
if type(showId) is not int:
|
||||||
|
raise ValueError(f"ShiftedSeasonController.findShiftedSeason(): Argument shiftedSeasonId is required to be of type int")
|
||||||
|
|
||||||
|
if type(originalSeason) is not int:
|
||||||
|
raise ValueError(f"ShiftedSeasonController.findShiftedSeason(): Argument originalSeason is required to be of type int")
|
||||||
|
|
||||||
|
if type(firstEpisode) is not int:
|
||||||
|
raise ValueError(f"ShiftedSeasonController.findShiftedSeason(): Argument firstEpisode is required to be of type int")
|
||||||
|
|
||||||
|
if type(lastEpisode) is not int:
|
||||||
|
raise ValueError(f"ShiftedSeasonController.findShiftedSeason(): Argument lastEpisode is required to be of type int")
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(ShiftedSeason).filter(ShiftedSeason.show_id == int(showId),
|
||||||
|
ShiftedSeason.original_season == int(originalSeason),
|
||||||
|
ShiftedSeason.first_episode == int(firstEpisode),
|
||||||
|
ShiftedSeason.last_episode == int(lastEpisode))
|
||||||
|
|
||||||
|
return q.first().getId() if q.count() else None
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"PatternController.findShiftedSeason(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
def getShiftedSeasonSiblings(self, showId: int):
|
||||||
|
|
||||||
|
if type(showId) is not int:
|
||||||
|
raise ValueError(f"ShiftedSeasonController.getShiftedSeasonSiblings(): Argument shiftedSeasonId is required to be of type int")
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(ShiftedSeason).filter(ShiftedSeason.show_id == int(showId))
|
||||||
|
|
||||||
|
return q.all()
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"PatternController.getShiftedSeasonSiblings(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def getShiftedSeason(self, shiftedSeasonId: int):
|
||||||
|
|
||||||
|
if type(shiftedSeasonId) is not int:
|
||||||
|
raise ValueError(f"ShiftedSeasonController.getShiftedSeason(): Argument shiftedSeasonId is required to be of type int")
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(ShiftedSeason).filter(ShiftedSeason.id == int(shiftedSeasonId))
|
||||||
|
|
||||||
|
return q.first() if q.count() else None
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"ShiftedSeasonController.getShiftedSeason(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def deleteShiftedSeason(self, shiftedSeasonId):
|
||||||
|
|
||||||
|
if type(shiftedSeasonId) is not int:
|
||||||
|
raise ValueError(f"ShiftedSeasonController.deleteShiftedSeason(): Argument shiftedSeasonId is required to be of type int")
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(ShiftedSeason).filter(ShiftedSeason.id == int(shiftedSeasonId))
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
|
||||||
|
#DAFUQ: https://stackoverflow.com/a/19245058
|
||||||
|
# q.delete()
|
||||||
|
shiftedSeason = q.first()
|
||||||
|
s.delete(shiftedSeason)
|
||||||
|
|
||||||
|
s.commit()
|
||||||
|
return True
|
||||||
|
return False
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"ShiftedSeasonController.deleteShiftedSeason(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def shiftSeason(self, showId, season, episode):
|
||||||
|
|
||||||
|
shiftedSeasonEntry: ShiftedSeason
|
||||||
|
for shiftedSeasonEntry in self.getShiftedSeasonSiblings(showId):
|
||||||
|
|
||||||
|
if (season == shiftedSeasonEntry.getOriginalSeason()
|
||||||
|
and (shiftedSeasonEntry.getFirstEpisode() == -1 or episode >= shiftedSeasonEntry.getFirstEpisode())
|
||||||
|
and (shiftedSeasonEntry.getLastEpisode() == -1 or episode <= shiftedSeasonEntry.getLastEpisode())):
|
||||||
|
|
||||||
|
shiftedSeason = season + shiftedSeasonEntry.getSeasonOffset()
|
||||||
|
shiftedEpisode = episode + shiftedSeasonEntry.getEpisodeOffset()
|
||||||
|
|
||||||
|
self.context['logger'].info(f"Shifting season: {season} episode: {episode} "
|
||||||
|
+f"-> season: {shiftedSeason} episode: {shiftedEpisode}")
|
||||||
|
|
||||||
|
return shiftedSeason, shiftedEpisode
|
||||||
|
|
||||||
|
return season, episode
|
||||||
125
src/ffx/shifted_season_delete_screen.py
Normal file
125
src/ffx/shifted_season_delete_screen.py
Normal file
@@ -0,0 +1,125 @@
|
|||||||
|
import click
|
||||||
|
|
||||||
|
from textual.screen import Screen
|
||||||
|
from textual.widgets import Header, Footer, Static, Button
|
||||||
|
from textual.containers import Grid
|
||||||
|
|
||||||
|
from .shifted_season_controller import ShiftedSeasonController
|
||||||
|
|
||||||
|
from ffx.model.shifted_season import ShiftedSeason
|
||||||
|
|
||||||
|
|
||||||
|
# Screen[dict[int, str, int]]
|
||||||
|
class ShiftedSeasonDeleteScreen(Screen):
|
||||||
|
|
||||||
|
CSS = """
|
||||||
|
|
||||||
|
Grid {
|
||||||
|
grid-size: 2;
|
||||||
|
grid-rows: 2 auto;
|
||||||
|
grid-columns: 30 330;
|
||||||
|
height: 100%;
|
||||||
|
width: 100%;
|
||||||
|
padding: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
Input {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
Button {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
#toplabel {
|
||||||
|
height: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
.two {
|
||||||
|
column-span: 2;
|
||||||
|
}
|
||||||
|
|
||||||
|
.box {
|
||||||
|
height: 100%;
|
||||||
|
border: solid green;
|
||||||
|
}
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, showId = None, shiftedSeasonId = None):
|
||||||
|
super().__init__()
|
||||||
|
|
||||||
|
self.context = self.app.getContext()
|
||||||
|
self.Session = self.context['database']['session'] # convenience
|
||||||
|
|
||||||
|
self.__ssc = ShiftedSeasonController(context = self.context)
|
||||||
|
|
||||||
|
self._showId = showId
|
||||||
|
self.__shiftedSeasonId = shiftedSeasonId
|
||||||
|
|
||||||
|
|
||||||
|
def on_mount(self):
|
||||||
|
|
||||||
|
shiftedSeason: ShiftedSeason = self.__ssc.getShiftedSeason(self.__shiftedSeasonId)
|
||||||
|
|
||||||
|
self.query_one("#static_show_id", Static).update(str(self._showId))
|
||||||
|
self.query_one("#static_original_season", Static).update(str(shiftedSeason.getOriginalSeason()))
|
||||||
|
self.query_one("#static_first_episode", Static).update(str(shiftedSeason.getFirstEpisode()))
|
||||||
|
self.query_one("#static_last_episode", Static).update(str(shiftedSeason.getLastEpisode()))
|
||||||
|
self.query_one("#static_season_offset", Static).update(str(shiftedSeason.getSeasonOffset()))
|
||||||
|
self.query_one("#static_episode_offset", Static).update(str(shiftedSeason.getEpisodeOffset()))
|
||||||
|
|
||||||
|
|
||||||
|
def compose(self):
|
||||||
|
|
||||||
|
yield Header()
|
||||||
|
|
||||||
|
with Grid():
|
||||||
|
|
||||||
|
yield Static("Are you sure to delete the following shifted season?", id="toplabel", classes="two")
|
||||||
|
|
||||||
|
yield Static(" ", classes="two")
|
||||||
|
|
||||||
|
yield Static("from show")
|
||||||
|
yield Static(" ", id="static_show_id")
|
||||||
|
|
||||||
|
yield Static(" ", classes="two")
|
||||||
|
|
||||||
|
yield Static("Original season")
|
||||||
|
yield Static(" ", id="static_original_season")
|
||||||
|
|
||||||
|
yield Static("First episode")
|
||||||
|
yield Static(" ", id="static_first_episode")
|
||||||
|
|
||||||
|
yield Static("Last episode")
|
||||||
|
yield Static(" ", id="static_last_episode")
|
||||||
|
|
||||||
|
yield Static("Season offset")
|
||||||
|
yield Static(" ", id="static_season_offset")
|
||||||
|
|
||||||
|
yield Static("Episode offset")
|
||||||
|
yield Static(" ", id="static_episode_offset")
|
||||||
|
|
||||||
|
yield Static(" ", classes="two")
|
||||||
|
|
||||||
|
yield Button("Delete", id="delete_button")
|
||||||
|
yield Button("Cancel", id="cancel_button")
|
||||||
|
|
||||||
|
yield Footer()
|
||||||
|
|
||||||
|
|
||||||
|
# Event handler for button press
|
||||||
|
def on_button_pressed(self, event: Button.Pressed) -> None:
|
||||||
|
|
||||||
|
if event.button.id == "delete_button":
|
||||||
|
|
||||||
|
if self.__shiftedSeasonId is None:
|
||||||
|
raise click.ClickException('ShiftedSeasonDeleteScreen.on_button_pressed(): shifted season id is undefined')
|
||||||
|
|
||||||
|
if self.__ssc.deleteShiftedSeason(self.__shiftedSeasonId):
|
||||||
|
self.dismiss(self.__shiftedSeasonId)
|
||||||
|
|
||||||
|
else:
|
||||||
|
#TODO: Meldung
|
||||||
|
self.app.pop_screen()
|
||||||
|
|
||||||
|
if event.button.id == "cancel_button":
|
||||||
|
self.app.pop_screen()
|
||||||
|
|
||||||
221
src/ffx/shifted_season_details_screen.py
Normal file
221
src/ffx/shifted_season_details_screen.py
Normal file
@@ -0,0 +1,221 @@
|
|||||||
|
from typing import List
|
||||||
|
|
||||||
|
from textual.screen import Screen
|
||||||
|
from textual.widgets import Header, Footer, Static, Button, Input
|
||||||
|
from textual.containers import Grid
|
||||||
|
|
||||||
|
from .shifted_season_controller import ShiftedSeasonController
|
||||||
|
|
||||||
|
from ffx.model.shifted_season import ShiftedSeason
|
||||||
|
|
||||||
|
|
||||||
|
# Screen[dict[int, str, int]]
|
||||||
|
class ShiftedSeasonDetailsScreen(Screen):
|
||||||
|
|
||||||
|
CSS = """
|
||||||
|
|
||||||
|
Grid {
|
||||||
|
grid-size: 3 10;
|
||||||
|
grid-rows: 2 2 2 2 2 2 2 2 2 2;
|
||||||
|
grid-columns: 40 40 40;
|
||||||
|
height: 100%;
|
||||||
|
width: 100%;
|
||||||
|
padding: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
Input {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
Button {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable {
|
||||||
|
min-height: 6;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable .datatable--cursor {
|
||||||
|
background: darkorange;
|
||||||
|
color: black;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable .datatable--header {
|
||||||
|
background: steelblue;
|
||||||
|
color: white;
|
||||||
|
}
|
||||||
|
|
||||||
|
#toplabel {
|
||||||
|
height: 1;
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
.two {
|
||||||
|
column-span: 3;
|
||||||
|
}
|
||||||
|
|
||||||
|
.three {
|
||||||
|
column-span: 3;
|
||||||
|
}
|
||||||
|
|
||||||
|
.four {
|
||||||
|
column-span: 4;
|
||||||
|
}
|
||||||
|
.five {
|
||||||
|
column-span: 5;
|
||||||
|
}
|
||||||
|
.six {
|
||||||
|
column-span: 6;
|
||||||
|
}
|
||||||
|
.seven {
|
||||||
|
column-span: 7;
|
||||||
|
}
|
||||||
|
|
||||||
|
.box {
|
||||||
|
height: 100%;
|
||||||
|
border: solid green;
|
||||||
|
}
|
||||||
|
|
||||||
|
.yellow {
|
||||||
|
tint: yellow 40%;
|
||||||
|
}
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, showId = None, shiftedSeasonId = None):
|
||||||
|
super().__init__()
|
||||||
|
|
||||||
|
self.context = self.app.getContext()
|
||||||
|
self.Session = self.context['database']['session'] # convenience
|
||||||
|
|
||||||
|
self.__ssc = ShiftedSeasonController(context = self.context)
|
||||||
|
|
||||||
|
self.__showId = showId
|
||||||
|
self.__shiftedSeasonId = shiftedSeasonId
|
||||||
|
|
||||||
|
def on_mount(self):
|
||||||
|
|
||||||
|
if self.__shiftedSeasonId is not None:
|
||||||
|
shiftedSeason: ShiftedSeason = self.__ssc.getShiftedSeason(self.__shiftedSeasonId)
|
||||||
|
|
||||||
|
originalSeason = shiftedSeason.getOriginalSeason()
|
||||||
|
self.query_one("#input_original_season", Input).value = str(originalSeason)
|
||||||
|
|
||||||
|
firstEpisode = shiftedSeason.getFirstEpisode()
|
||||||
|
self.query_one("#input_first_episode", Input).value = str(firstEpisode) if firstEpisode != -1 else ''
|
||||||
|
|
||||||
|
lastEpisode = shiftedSeason.getLastEpisode()
|
||||||
|
self.query_one("#input_last_episode", Input).value = str(lastEpisode) if lastEpisode != -1 else ''
|
||||||
|
|
||||||
|
seasonOffset = shiftedSeason.getSeasonOffset()
|
||||||
|
self.query_one("#input_season_offset", Input).value = str(seasonOffset) if seasonOffset else ''
|
||||||
|
|
||||||
|
episodeOffset = shiftedSeason.getEpisodeOffset()
|
||||||
|
self.query_one("#input_episode_offset", Input).value = str(episodeOffset) if episodeOffset else ''
|
||||||
|
|
||||||
|
|
||||||
|
def compose(self):
|
||||||
|
|
||||||
|
yield Header()
|
||||||
|
|
||||||
|
with Grid():
|
||||||
|
|
||||||
|
# 1
|
||||||
|
yield Static("Edit shifted season" if self.__shiftedSeasonId is not None else "New shifted season", id="toplabel", classes="three")
|
||||||
|
|
||||||
|
# 2
|
||||||
|
yield Static(" ", classes="three")
|
||||||
|
|
||||||
|
# 3
|
||||||
|
yield Static("Original season")
|
||||||
|
yield Input(id="input_original_season", classes="two")
|
||||||
|
|
||||||
|
# 4
|
||||||
|
yield Static("First Episode")
|
||||||
|
yield Input(id="input_first_episode", classes="two")
|
||||||
|
|
||||||
|
# 5
|
||||||
|
yield Static("Last Episode")
|
||||||
|
yield Input(id="input_last_episode", classes="two")
|
||||||
|
|
||||||
|
# 6
|
||||||
|
yield Static("Season offset")
|
||||||
|
yield Input(id="input_season_offset", classes="two")
|
||||||
|
|
||||||
|
# 7
|
||||||
|
yield Static("Episode offset")
|
||||||
|
yield Input(id="input_episode_offset", classes="two")
|
||||||
|
|
||||||
|
# 8
|
||||||
|
yield Static(" ", classes="three")
|
||||||
|
|
||||||
|
# 9
|
||||||
|
yield Button("Save", id="save_button")
|
||||||
|
yield Button("Cancel", id="cancel_button")
|
||||||
|
yield Static(" ")
|
||||||
|
|
||||||
|
# 10
|
||||||
|
yield Static(" ", classes="three")
|
||||||
|
|
||||||
|
yield Footer()
|
||||||
|
|
||||||
|
|
||||||
|
def getShiftedSeasonObjFromInput(self):
|
||||||
|
|
||||||
|
shiftedSeasonObj = {}
|
||||||
|
|
||||||
|
originalSeason = self.query_one("#input_original_season", Input).value
|
||||||
|
if not originalSeason:
|
||||||
|
return None
|
||||||
|
shiftedSeasonObj['original_season'] = int(originalSeason)
|
||||||
|
|
||||||
|
try:
|
||||||
|
shiftedSeasonObj['first_episode'] = int(self.query_one("#input_first_episode", Input).value)
|
||||||
|
except ValueError:
|
||||||
|
shiftedSeasonObj['first_episode'] = -1
|
||||||
|
|
||||||
|
try:
|
||||||
|
shiftedSeasonObj['last_episode'] = int(self.query_one("#input_last_episode", Input).value)
|
||||||
|
except ValueError:
|
||||||
|
shiftedSeasonObj['last_episode'] = -1
|
||||||
|
|
||||||
|
try:
|
||||||
|
shiftedSeasonObj['season_offset'] = int(self.query_one("#input_season_offset", Input).value)
|
||||||
|
except ValueError:
|
||||||
|
shiftedSeasonObj['season_offset'] = 0
|
||||||
|
|
||||||
|
try:
|
||||||
|
shiftedSeasonObj['episode_offset'] = int(self.query_one("#input_episode_offset", Input).value)
|
||||||
|
except ValueError:
|
||||||
|
shiftedSeasonObj['episode_offset'] = 0
|
||||||
|
|
||||||
|
return shiftedSeasonObj
|
||||||
|
|
||||||
|
|
||||||
|
# Event handler for button press
|
||||||
|
def on_button_pressed(self, event: Button.Pressed) -> None:
|
||||||
|
|
||||||
|
# Check if the button pressed is the one we are interested in
|
||||||
|
if event.button.id == "save_button":
|
||||||
|
|
||||||
|
shiftedSeasonObj = self.getShiftedSeasonObjFromInput()
|
||||||
|
|
||||||
|
if shiftedSeasonObj is not None:
|
||||||
|
|
||||||
|
if self.__shiftedSeasonId is not None:
|
||||||
|
|
||||||
|
if self.__ssc.checkShiftedSeason(self.__showId, shiftedSeasonObj,
|
||||||
|
shiftedSeasonId = self.__shiftedSeasonId):
|
||||||
|
if self.__ssc.updateShiftedSeason(self.__shiftedSeasonId, shiftedSeasonObj):
|
||||||
|
self.dismiss((self.__shiftedSeasonId, shiftedSeasonObj))
|
||||||
|
else:
|
||||||
|
#TODO: Meldung
|
||||||
|
self.app.pop_screen()
|
||||||
|
|
||||||
|
else:
|
||||||
|
if self.__ssc.checkShiftedSeason(self.__showId, shiftedSeasonObj):
|
||||||
|
self.__shiftedSeasonId = self.__ssc.addShiftedSeason(self.__showId, shiftedSeasonObj)
|
||||||
|
self.dismiss((self.__shiftedSeasonId, shiftedSeasonObj))
|
||||||
|
|
||||||
|
|
||||||
|
if event.button.id == "cancel_button":
|
||||||
|
self.app.pop_screen()
|
||||||
133
src/ffx/show_controller.py
Normal file
133
src/ffx/show_controller.py
Normal file
@@ -0,0 +1,133 @@
|
|||||||
|
import click
|
||||||
|
|
||||||
|
from ffx.model.show import Show
|
||||||
|
from ffx.show_descriptor import ShowDescriptor
|
||||||
|
|
||||||
|
|
||||||
|
class ShowController():
|
||||||
|
|
||||||
|
def __init__(self, context):
|
||||||
|
|
||||||
|
self.context = context
|
||||||
|
self.Session = self.context['database']['session'] # convenience
|
||||||
|
|
||||||
|
|
||||||
|
def getShowDescriptor(self, showId):
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Show).filter(Show.id == showId)
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
show: Show = q.first()
|
||||||
|
return show.getDescriptor(self.context)
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"ShowController.getShowDescriptor(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
def getShow(self, showId):
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Show).filter(Show.id == showId)
|
||||||
|
|
||||||
|
return q.first() if q.count() else None
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"ShowController.getShow(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
def getAllShows(self):
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Show)
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
return q.all()
|
||||||
|
else:
|
||||||
|
return []
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"ShowController.getAllShows(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def updateShow(self, showDescriptor: ShowDescriptor):
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Show).filter(Show.id == showDescriptor.getId())
|
||||||
|
|
||||||
|
if not q.count():
|
||||||
|
show = Show(id = int(showDescriptor.getId()),
|
||||||
|
name = str(showDescriptor.getName()),
|
||||||
|
year = int(showDescriptor.getYear()),
|
||||||
|
index_season_digits = showDescriptor.getIndexSeasonDigits(),
|
||||||
|
index_episode_digits = showDescriptor.getIndexEpisodeDigits(),
|
||||||
|
indicator_season_digits = showDescriptor.getIndicatorSeasonDigits(),
|
||||||
|
indicator_episode_digits = showDescriptor.getIndicatorEpisodeDigits())
|
||||||
|
|
||||||
|
s.add(show)
|
||||||
|
s.commit()
|
||||||
|
return True
|
||||||
|
else:
|
||||||
|
|
||||||
|
currentShow = q.first()
|
||||||
|
|
||||||
|
changed = False
|
||||||
|
if currentShow.name != str(showDescriptor.getName()):
|
||||||
|
currentShow.name = str(showDescriptor.getName())
|
||||||
|
changed = True
|
||||||
|
if currentShow.year != int(showDescriptor.getYear()):
|
||||||
|
currentShow.year = int(showDescriptor.getYear())
|
||||||
|
changed = True
|
||||||
|
|
||||||
|
if currentShow.index_season_digits != int(showDescriptor.getIndexSeasonDigits()):
|
||||||
|
currentShow.index_season_digits = int(showDescriptor.getIndexSeasonDigits())
|
||||||
|
changed = True
|
||||||
|
if currentShow.index_episode_digits != int(showDescriptor.getIndexEpisodeDigits()):
|
||||||
|
currentShow.index_episode_digits = int(showDescriptor.getIndexEpisodeDigits())
|
||||||
|
changed = True
|
||||||
|
if currentShow.indicator_season_digits != int(showDescriptor.getIndicatorSeasonDigits()):
|
||||||
|
currentShow.indicator_season_digits = int(showDescriptor.getIndicatorSeasonDigits())
|
||||||
|
changed = True
|
||||||
|
if currentShow.indicator_episode_digits != int(showDescriptor.getIndicatorEpisodeDigits()):
|
||||||
|
currentShow.indicator_episode_digits = int(showDescriptor.getIndicatorEpisodeDigits())
|
||||||
|
changed = True
|
||||||
|
|
||||||
|
if changed:
|
||||||
|
s.commit()
|
||||||
|
return changed
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"ShowController.updateShow(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def deleteShow(self, show_id):
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Show).filter(Show.id == int(show_id))
|
||||||
|
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
|
||||||
|
#DAFUQ: https://stackoverflow.com/a/19245058
|
||||||
|
# q.delete()
|
||||||
|
show = q.first()
|
||||||
|
s.delete(show)
|
||||||
|
|
||||||
|
s.commit()
|
||||||
|
return True
|
||||||
|
return False
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"ShowController.deleteShow(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
95
src/ffx/show_delete_screen.py
Normal file
95
src/ffx/show_delete_screen.py
Normal file
@@ -0,0 +1,95 @@
|
|||||||
|
from textual.screen import Screen
|
||||||
|
from textual.widgets import Header, Footer, Static, Button
|
||||||
|
from textual.containers import Grid
|
||||||
|
|
||||||
|
from .show_controller import ShowController
|
||||||
|
|
||||||
|
# Screen[dict[int, str, int]]
|
||||||
|
class ShowDeleteScreen(Screen):
|
||||||
|
|
||||||
|
CSS = """
|
||||||
|
|
||||||
|
Grid {
|
||||||
|
grid-size: 2;
|
||||||
|
grid-rows: 2 auto;
|
||||||
|
grid-columns: 30 auto;
|
||||||
|
height: 100%;
|
||||||
|
width: 100%;
|
||||||
|
padding: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
Input {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
Button {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
#toplabel {
|
||||||
|
height: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
.two {
|
||||||
|
column-span: 2;
|
||||||
|
}
|
||||||
|
|
||||||
|
.box {
|
||||||
|
height: 100%;
|
||||||
|
border: solid green;
|
||||||
|
}
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, showId = None):
|
||||||
|
super().__init__()
|
||||||
|
|
||||||
|
self.context = self.app.getContext()
|
||||||
|
self.Session = self.context['database']['session'] # convenience
|
||||||
|
|
||||||
|
self.__sc = ShowController(context = self.context)
|
||||||
|
|
||||||
|
self.__showDescriptor = self.__sc.getShowDescriptor(showId) if showId is not None else {}
|
||||||
|
|
||||||
|
|
||||||
|
def on_mount(self):
|
||||||
|
if not self.__showDescriptor is None:
|
||||||
|
self.query_one("#showlabel", Static).update(f"{self.__showDescriptor.getId()} - {self.__showDescriptor.getName()} ({self.__showDescriptor.getYear()})")
|
||||||
|
|
||||||
|
|
||||||
|
def compose(self):
|
||||||
|
|
||||||
|
yield Header()
|
||||||
|
|
||||||
|
with Grid():
|
||||||
|
|
||||||
|
yield Static("Are you sure to delete the following show?", id="toplabel", classes="two")
|
||||||
|
|
||||||
|
yield Static("", classes="two")
|
||||||
|
|
||||||
|
yield Static("", id="showlabel")
|
||||||
|
yield Static("")
|
||||||
|
|
||||||
|
yield Static("", classes="two")
|
||||||
|
|
||||||
|
yield Static("", classes="two")
|
||||||
|
|
||||||
|
yield Button("Delete", id="delete_button")
|
||||||
|
yield Button("Cancel", id="cancel_button")
|
||||||
|
|
||||||
|
|
||||||
|
yield Footer()
|
||||||
|
|
||||||
|
|
||||||
|
# Event handler for button press
|
||||||
|
def on_button_pressed(self, event: Button.Pressed) -> None:
|
||||||
|
|
||||||
|
if event.button.id == "delete_button":
|
||||||
|
|
||||||
|
if not self.__showDescriptor is None:
|
||||||
|
if self.__sc.deleteShow(self.__showDescriptor.getId()):
|
||||||
|
self.dismiss(self.__showDescriptor)
|
||||||
|
|
||||||
|
else:
|
||||||
|
#TODO: Meldung
|
||||||
|
self.app.pop_screen()
|
||||||
|
|
||||||
|
if event.button.id == "cancel_button":
|
||||||
|
self.app.pop_screen()
|
||||||
106
src/ffx/show_descriptor.py
Normal file
106
src/ffx/show_descriptor.py
Normal file
@@ -0,0 +1,106 @@
|
|||||||
|
import logging
|
||||||
|
|
||||||
|
|
||||||
|
class ShowDescriptor():
|
||||||
|
"""This class represents the structural content of a media file including streams and metadata"""
|
||||||
|
|
||||||
|
CONTEXT_KEY = 'context'
|
||||||
|
|
||||||
|
ID_KEY = 'id'
|
||||||
|
NAME_KEY = 'name'
|
||||||
|
YEAR_KEY = 'year'
|
||||||
|
|
||||||
|
INDEX_SEASON_DIGITS_KEY = 'index_season_digits'
|
||||||
|
INDEX_EPISODE_DIGITS_KEY = 'index_episode_digits'
|
||||||
|
INDICATOR_SEASON_DIGITS_KEY = 'indicator_season_digits'
|
||||||
|
INDICATOR_EPISODE_DIGITS_KEY = 'indicator_episode_digits'
|
||||||
|
|
||||||
|
DEFAULT_INDEX_SEASON_DIGITS = 2
|
||||||
|
DEFAULT_INDEX_EPISODE_DIGITS = 2
|
||||||
|
DEFAULT_INDICATOR_SEASON_DIGITS = 2
|
||||||
|
DEFAULT_INDICATOR_EPISODE_DIGITS = 2
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, **kwargs):
|
||||||
|
|
||||||
|
if ShowDescriptor.CONTEXT_KEY in kwargs.keys():
|
||||||
|
if type(kwargs[ShowDescriptor.CONTEXT_KEY]) is not dict:
|
||||||
|
raise TypeError(
|
||||||
|
f"ShowDescriptor.__init__(): Argument {ShowDescriptor.CONTEXT_KEY} is required to be of type dict"
|
||||||
|
)
|
||||||
|
self.__context = kwargs[ShowDescriptor.CONTEXT_KEY]
|
||||||
|
self.__logger = self.__context['logger']
|
||||||
|
else:
|
||||||
|
self.__context = {}
|
||||||
|
self.__logger = logging.getLogger('FFX')
|
||||||
|
self.__logger.addHandler(logging.NullHandler())
|
||||||
|
|
||||||
|
if ShowDescriptor.ID_KEY in kwargs.keys():
|
||||||
|
if type(kwargs[ShowDescriptor.ID_KEY]) is not int:
|
||||||
|
raise TypeError(f"ShowDescriptor.__init__(): Argument {ShowDescriptor.ID_KEY} is required to be of type int")
|
||||||
|
self.__showId = kwargs[ShowDescriptor.ID_KEY]
|
||||||
|
else:
|
||||||
|
self.__showId = -1
|
||||||
|
|
||||||
|
if ShowDescriptor.NAME_KEY in kwargs.keys():
|
||||||
|
if type(kwargs[ShowDescriptor.NAME_KEY]) is not str:
|
||||||
|
raise TypeError(f"ShowDescriptor.__init__(): Argument {ShowDescriptor.NAME_KEY} is required to be of type str")
|
||||||
|
self.__showName = kwargs[ShowDescriptor.NAME_KEY]
|
||||||
|
else:
|
||||||
|
self.__showName = ''
|
||||||
|
|
||||||
|
if ShowDescriptor.YEAR_KEY in kwargs.keys():
|
||||||
|
if type(kwargs[ShowDescriptor.YEAR_KEY]) is not int:
|
||||||
|
raise TypeError(f"ShowDescriptor.__init__(): Argument {ShowDescriptor.YEAR_KEY} is required to be of type int")
|
||||||
|
self.__showYear = kwargs[ShowDescriptor.YEAR_KEY]
|
||||||
|
else:
|
||||||
|
self.__showYear = -1
|
||||||
|
|
||||||
|
|
||||||
|
if ShowDescriptor.INDEX_SEASON_DIGITS_KEY in kwargs.keys():
|
||||||
|
if type(kwargs[ShowDescriptor.INDEX_SEASON_DIGITS_KEY]) is not int:
|
||||||
|
raise TypeError(f"ShowDescriptor.__init__(): Argument {ShowDescriptor.INDEX_SEASON_DIGITS_KEY} is required to be of type int")
|
||||||
|
self.__indexSeasonDigits = kwargs[ShowDescriptor.INDEX_SEASON_DIGITS_KEY]
|
||||||
|
else:
|
||||||
|
self.__indexSeasonDigits = ShowDescriptor.DEFAULT_INDEX_SEASON_DIGITS
|
||||||
|
|
||||||
|
if ShowDescriptor.INDEX_EPISODE_DIGITS_KEY in kwargs.keys():
|
||||||
|
if type(kwargs[ShowDescriptor.INDEX_EPISODE_DIGITS_KEY]) is not int:
|
||||||
|
raise TypeError(f"ShowDescriptor.__init__(): Argument {ShowDescriptor.INDEX_EPISODE_DIGITS_KEY} is required to be of type int")
|
||||||
|
self.__indexEpisodeDigits = kwargs[ShowDescriptor.INDEX_EPISODE_DIGITS_KEY]
|
||||||
|
else:
|
||||||
|
self.__indexEpisodeDigits = ShowDescriptor.DEFAULT_INDEX_EPISODE_DIGITS
|
||||||
|
|
||||||
|
if ShowDescriptor.INDICATOR_SEASON_DIGITS_KEY in kwargs.keys():
|
||||||
|
if type(kwargs[ShowDescriptor.INDICATOR_SEASON_DIGITS_KEY]) is not int:
|
||||||
|
raise TypeError(f"ShowDescriptor.__init__(): Argument {ShowDescriptor.INDICATOR_SEASON_DIGITS_KEY} is required to be of type int")
|
||||||
|
self.__indicatorSeasonDigits = kwargs[ShowDescriptor.INDICATOR_SEASON_DIGITS_KEY]
|
||||||
|
else:
|
||||||
|
self.__indicatorSeasonDigits = ShowDescriptor.DEFAULT_INDICATOR_SEASON_DIGITS
|
||||||
|
|
||||||
|
if ShowDescriptor.INDICATOR_EPISODE_DIGITS_KEY in kwargs.keys():
|
||||||
|
if type(kwargs[ShowDescriptor.INDICATOR_EPISODE_DIGITS_KEY]) is not int:
|
||||||
|
raise TypeError(f"ShowDescriptor.__init__(): Argument {ShowDescriptor.INDICATOR_EPISODE_DIGITS_KEY} is required to be of type int")
|
||||||
|
self.__indicatorEpisodeDigits = kwargs[ShowDescriptor.INDICATOR_EPISODE_DIGITS_KEY]
|
||||||
|
else:
|
||||||
|
self.__indicatorEpisodeDigits = ShowDescriptor.DEFAULT_INDICATOR_EPISODE_DIGITS
|
||||||
|
|
||||||
|
|
||||||
|
def getId(self):
|
||||||
|
return self.__showId
|
||||||
|
def getName(self):
|
||||||
|
return self.__showName
|
||||||
|
def getYear(self):
|
||||||
|
return self.__showYear
|
||||||
|
|
||||||
|
def getIndexSeasonDigits(self):
|
||||||
|
return self.__indexSeasonDigits
|
||||||
|
def getIndexEpisodeDigits(self):
|
||||||
|
return self.__indexEpisodeDigits
|
||||||
|
def getIndicatorSeasonDigits(self):
|
||||||
|
return self.__indicatorSeasonDigits
|
||||||
|
def getIndicatorEpisodeDigits(self):
|
||||||
|
return self.__indicatorEpisodeDigits
|
||||||
|
|
||||||
|
def getFilenamePrefix(self):
|
||||||
|
return f"{self.__showName} ({str(self.__showYear)})"
|
||||||
492
src/ffx/show_details_screen.py
Normal file
492
src/ffx/show_details_screen.py
Normal file
@@ -0,0 +1,492 @@
|
|||||||
|
import click
|
||||||
|
|
||||||
|
from textual.screen import Screen
|
||||||
|
from textual.widgets import Header, Footer, Static, Button, DataTable, Input
|
||||||
|
from textual.containers import Grid
|
||||||
|
from textual.widgets._data_table import CellDoesNotExist
|
||||||
|
|
||||||
|
from ffx.model.pattern import Pattern
|
||||||
|
|
||||||
|
from .pattern_details_screen import PatternDetailsScreen
|
||||||
|
from .pattern_delete_screen import PatternDeleteScreen
|
||||||
|
|
||||||
|
from .show_controller import ShowController
|
||||||
|
from .pattern_controller import PatternController
|
||||||
|
from .tmdb_controller import TmdbController
|
||||||
|
from .shifted_season_controller import ShiftedSeasonController
|
||||||
|
|
||||||
|
from .show_descriptor import ShowDescriptor
|
||||||
|
|
||||||
|
from .shifted_season_details_screen import ShiftedSeasonDetailsScreen
|
||||||
|
from .shifted_season_delete_screen import ShiftedSeasonDeleteScreen
|
||||||
|
|
||||||
|
from ffx.model.shifted_season import ShiftedSeason
|
||||||
|
|
||||||
|
from .helper import filterFilename
|
||||||
|
|
||||||
|
|
||||||
|
# Screen[dict[int, str, int]]
|
||||||
|
class ShowDetailsScreen(Screen):
|
||||||
|
|
||||||
|
CSS = """
|
||||||
|
|
||||||
|
Grid {
|
||||||
|
grid-size: 5 16;
|
||||||
|
grid-rows: 2 2 2 2 2 2 2 2 2 2 2 9 2 9 2 2;
|
||||||
|
grid-columns: 30 30 30 30 30;
|
||||||
|
height: 100%;
|
||||||
|
width: 100%;
|
||||||
|
padding: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
Input {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
Button {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable {
|
||||||
|
column-span: 2;
|
||||||
|
min-height: 8;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable .datatable--cursor {
|
||||||
|
background: darkorange;
|
||||||
|
color: black;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable .datatable--header {
|
||||||
|
background: steelblue;
|
||||||
|
color: white;
|
||||||
|
}
|
||||||
|
|
||||||
|
#toplabel {
|
||||||
|
height: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
.two {
|
||||||
|
column-span: 2;
|
||||||
|
}
|
||||||
|
.three {
|
||||||
|
column-span: 3;
|
||||||
|
}
|
||||||
|
.four {
|
||||||
|
column-span: 4;
|
||||||
|
}
|
||||||
|
.five {
|
||||||
|
column-span: 5;
|
||||||
|
}
|
||||||
|
|
||||||
|
.box {
|
||||||
|
height: 100%;
|
||||||
|
border: solid green;
|
||||||
|
}
|
||||||
|
"""
|
||||||
|
|
||||||
|
BINDINGS = [
|
||||||
|
("a", "add_pattern", "Add Pattern"),
|
||||||
|
("e", "edit_pattern", "Edit Pattern"),
|
||||||
|
("r", "remove_pattern", "Remove Pattern"),
|
||||||
|
]
|
||||||
|
|
||||||
|
def __init__(self, showId = None):
|
||||||
|
super().__init__()
|
||||||
|
|
||||||
|
self.context = self.app.getContext()
|
||||||
|
self.Session = self.context['database']['session'] # convenience
|
||||||
|
|
||||||
|
self.__sc = ShowController(context = self.context)
|
||||||
|
self.__pc = PatternController(context = self.context)
|
||||||
|
self.__tc = TmdbController()
|
||||||
|
self.__ssc = ShiftedSeasonController(context = self.context)
|
||||||
|
|
||||||
|
self.__showDescriptor = self.__sc.getShowDescriptor(showId) if showId is not None else None
|
||||||
|
|
||||||
|
|
||||||
|
def loadPatterns(self, show_id : int):
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Pattern).filter(Pattern.show_id == int(show_id))
|
||||||
|
|
||||||
|
return [{'id': int(p.id), 'pattern': str(p.pattern)} for p in q.all()]
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"ShowDetailsScreen.loadPatterns(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def updateShiftedSeasons(self):
|
||||||
|
|
||||||
|
self.shiftedSeasonsTable.clear()
|
||||||
|
|
||||||
|
if not self.__showDescriptor is None:
|
||||||
|
|
||||||
|
showId = int(self.__showDescriptor.getId())
|
||||||
|
|
||||||
|
shiftedSeason: ShiftedSeason
|
||||||
|
for shiftedSeason in self.__ssc.getShiftedSeasonSiblings(showId=showId):
|
||||||
|
|
||||||
|
shiftedSeasonObj = shiftedSeason.getObj()
|
||||||
|
|
||||||
|
firstEpisode = shiftedSeasonObj['first_episode']
|
||||||
|
firstEpisodeStr = str(firstEpisode) if firstEpisode != -1 else ''
|
||||||
|
|
||||||
|
lastEpisode = shiftedSeasonObj['last_episode']
|
||||||
|
lastEpisodeStr = str(lastEpisode) if lastEpisode != -1 else ''
|
||||||
|
|
||||||
|
row = (shiftedSeasonObj['original_season'],
|
||||||
|
firstEpisodeStr,
|
||||||
|
lastEpisodeStr,
|
||||||
|
shiftedSeasonObj['season_offset'],
|
||||||
|
shiftedSeasonObj['episode_offset'])
|
||||||
|
|
||||||
|
self.shiftedSeasonsTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def on_mount(self):
|
||||||
|
|
||||||
|
if self.__showDescriptor is not None:
|
||||||
|
|
||||||
|
showId = int(self.__showDescriptor.getId())
|
||||||
|
|
||||||
|
self.query_one("#id_static", Static).update(str(showId))
|
||||||
|
self.query_one("#name_input", Input).value = str(self.__showDescriptor.getName())
|
||||||
|
self.query_one("#year_input", Input).value = str(self.__showDescriptor.getYear())
|
||||||
|
|
||||||
|
self.query_one("#index_season_digits_input", Input).value = str(self.__showDescriptor.getIndexSeasonDigits())
|
||||||
|
self.query_one("#index_episode_digits_input", Input).value = str(self.__showDescriptor.getIndexEpisodeDigits())
|
||||||
|
self.query_one("#indicator_season_digits_input", Input).value = str(self.__showDescriptor.getIndicatorSeasonDigits())
|
||||||
|
self.query_one("#indicator_episode_digits_input", Input).value = str(self.__showDescriptor.getIndicatorEpisodeDigits())
|
||||||
|
|
||||||
|
|
||||||
|
#raise click.ClickException(f"show_id {showId}")
|
||||||
|
patternList = self.loadPatterns(showId)
|
||||||
|
# raise click.ClickException(f"patternList {patternList}")
|
||||||
|
for pattern in patternList:
|
||||||
|
row = (pattern['pattern'],)
|
||||||
|
self.patternTable.add_row(*map(str, row))
|
||||||
|
|
||||||
|
self.updateShiftedSeasons()
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
self.query_one("#index_season_digits_input", Input).value = "2"
|
||||||
|
self.query_one("#index_episode_digits_input", Input).value = "2"
|
||||||
|
self.query_one("#indicator_season_digits_input", Input).value = "2"
|
||||||
|
self.query_one("#indicator_episode_digits_input", Input).value = "2"
|
||||||
|
|
||||||
|
|
||||||
|
def getSelectedPatternDescriptor(self):
|
||||||
|
|
||||||
|
selectedPattern = {}
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
# Fetch the currently selected row when 'Enter' is pressed
|
||||||
|
#selected_row_index = self.table.cursor_row
|
||||||
|
row_key, col_key = self.patternTable.coordinate_to_cell_key(self.patternTable.cursor_coordinate)
|
||||||
|
|
||||||
|
if row_key is not None:
|
||||||
|
selected_row_data = self.patternTable.get_row(row_key)
|
||||||
|
|
||||||
|
selectedPattern['show_id'] = self.__showDescriptor.getId()
|
||||||
|
selectedPattern['pattern'] = str(selected_row_data[0])
|
||||||
|
|
||||||
|
except CellDoesNotExist:
|
||||||
|
pass
|
||||||
|
|
||||||
|
return selectedPattern
|
||||||
|
|
||||||
|
|
||||||
|
def getSelectedShiftedSeasonObjFromInput(self):
|
||||||
|
|
||||||
|
shiftedSeasonObj = {}
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
# Fetch the currently selected row when 'Enter' is pressed
|
||||||
|
#selected_row_index = self.table.cursor_row
|
||||||
|
row_key, col_key = self.shiftedSeasonsTable.coordinate_to_cell_key(self.shiftedSeasonsTable.cursor_coordinate)
|
||||||
|
|
||||||
|
if row_key is not None:
|
||||||
|
selected_row_data = self.shiftedSeasonsTable.get_row(row_key)
|
||||||
|
|
||||||
|
shiftedSeasonObj['original_season'] = int(selected_row_data[0])
|
||||||
|
shiftedSeasonObj['first_episode'] = int(selected_row_data[1]) if selected_row_data[1].isnumeric() else -1
|
||||||
|
shiftedSeasonObj['last_episode'] = int(selected_row_data[2]) if selected_row_data[2].isnumeric() else -1
|
||||||
|
shiftedSeasonObj['season_offset'] = int(selected_row_data[3]) if selected_row_data[3].isnumeric() else 0
|
||||||
|
shiftedSeasonObj['episode_offset'] = int(selected_row_data[4]) if selected_row_data[4].isnumeric() else 0
|
||||||
|
|
||||||
|
|
||||||
|
if self.__showDescriptor is not None:
|
||||||
|
|
||||||
|
showId = int(self.__showDescriptor.getId())
|
||||||
|
|
||||||
|
shiftedSeasonId = self.__ssc.findShiftedSeason(showId,
|
||||||
|
originalSeason=shiftedSeasonObj['original_season'],
|
||||||
|
firstEpisode=shiftedSeasonObj['first_episode'],
|
||||||
|
lastEpisode=shiftedSeasonObj['last_episode'])
|
||||||
|
if shiftedSeasonId is not None:
|
||||||
|
shiftedSeasonObj['id'] = shiftedSeasonId
|
||||||
|
|
||||||
|
except CellDoesNotExist:
|
||||||
|
pass
|
||||||
|
|
||||||
|
return shiftedSeasonObj
|
||||||
|
|
||||||
|
|
||||||
|
def action_add_pattern(self):
|
||||||
|
if not self.__showDescriptor is None:
|
||||||
|
self.app.push_screen(PatternDetailsScreen(showId = self.__showDescriptor.getId()), self.handle_add_pattern)
|
||||||
|
|
||||||
|
|
||||||
|
def handle_add_pattern(self, screenResult):
|
||||||
|
|
||||||
|
pattern = (screenResult['pattern'],)
|
||||||
|
self.patternTable.add_row(*map(str, pattern))
|
||||||
|
|
||||||
|
|
||||||
|
def action_edit_pattern(self):
|
||||||
|
|
||||||
|
selectedPatternDescriptor = self.getSelectedPatternDescriptor()
|
||||||
|
|
||||||
|
if selectedPatternDescriptor:
|
||||||
|
|
||||||
|
selectedPatternId = self.__pc.findPattern(selectedPatternDescriptor)
|
||||||
|
|
||||||
|
if selectedPatternId is None:
|
||||||
|
raise click.ClickException(f"ShowDetailsScreen.action_edit_pattern(): Pattern to edit has no id")
|
||||||
|
|
||||||
|
self.app.push_screen(PatternDetailsScreen(patternId = selectedPatternId, showId = self.__showDescriptor.getId()), self.handle_edit_pattern) # <-
|
||||||
|
|
||||||
|
|
||||||
|
def handle_edit_pattern(self, screenResult):
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
row_key, col_key = self.patternTable.coordinate_to_cell_key(self.patternTable.cursor_coordinate)
|
||||||
|
self.patternTable.update_cell(row_key, self.column_key_pattern, screenResult['pattern'])
|
||||||
|
|
||||||
|
except CellDoesNotExist:
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
def action_remove_pattern(self):
|
||||||
|
|
||||||
|
selectedPatternDescriptor = self.getSelectedPatternDescriptor()
|
||||||
|
|
||||||
|
if selectedPatternDescriptor:
|
||||||
|
|
||||||
|
selectedPatternId = self.__pc.findPattern(selectedPatternDescriptor)
|
||||||
|
|
||||||
|
if selectedPatternId is None:
|
||||||
|
raise click.ClickException(f"ShowDetailsScreen.action_remove_pattern(): Pattern to remove has no id")
|
||||||
|
|
||||||
|
self.app.push_screen(PatternDeleteScreen(patternId = selectedPatternId, showId = self.__showDescriptor.getId()), self.handle_remove_pattern)
|
||||||
|
|
||||||
|
|
||||||
|
def handle_remove_pattern(self, pattern):
|
||||||
|
|
||||||
|
try:
|
||||||
|
row_key, col_key = self.patternTable.coordinate_to_cell_key(self.patternTable.cursor_coordinate)
|
||||||
|
self.patternTable.remove_row(row_key)
|
||||||
|
|
||||||
|
except CellDoesNotExist:
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
def compose(self):
|
||||||
|
|
||||||
|
# Create the DataTable widget
|
||||||
|
self.patternTable = DataTable(classes="five")
|
||||||
|
|
||||||
|
# Define the columns with headers
|
||||||
|
self.column_key_pattern = self.patternTable.add_column("Pattern", width=150)
|
||||||
|
|
||||||
|
self.patternTable.cursor_type = 'row'
|
||||||
|
|
||||||
|
|
||||||
|
self.shiftedSeasonsTable = DataTable(classes="five")
|
||||||
|
|
||||||
|
self.column_key_original_season = self.shiftedSeasonsTable.add_column("Original Season", width=30)
|
||||||
|
self.column_key_first_episode = self.shiftedSeasonsTable.add_column("First Episode", width=30)
|
||||||
|
self.column_key_last_episode = self.shiftedSeasonsTable.add_column("Last Episode", width=30)
|
||||||
|
self.column_key_season_offset = self.shiftedSeasonsTable.add_column("Season Offset", width=30)
|
||||||
|
self.column_key_episode_offset = self.shiftedSeasonsTable.add_column("Episode Offset", width=30)
|
||||||
|
|
||||||
|
self.shiftedSeasonsTable.cursor_type = 'row'
|
||||||
|
|
||||||
|
|
||||||
|
yield Header()
|
||||||
|
|
||||||
|
with Grid():
|
||||||
|
|
||||||
|
# 1
|
||||||
|
yield Static("Show" if not self.__showDescriptor is None else "New Show", id="toplabel")
|
||||||
|
yield Button("Identify", id="identify_button")
|
||||||
|
yield Static(" ", classes="three")
|
||||||
|
|
||||||
|
# 2
|
||||||
|
yield Static("ID")
|
||||||
|
if not self.__showDescriptor is None:
|
||||||
|
yield Static("", id="id_static", classes="four")
|
||||||
|
else:
|
||||||
|
yield Input(type="integer", id="id_input", classes="four")
|
||||||
|
|
||||||
|
# 3
|
||||||
|
yield Static("Name")
|
||||||
|
yield Input(type="text", id="name_input", classes="four")
|
||||||
|
|
||||||
|
# 4
|
||||||
|
yield Static("Year")
|
||||||
|
yield Input(type="integer", id="year_input", classes="four")
|
||||||
|
|
||||||
|
#5
|
||||||
|
yield Static(" ", classes="five")
|
||||||
|
|
||||||
|
#6
|
||||||
|
yield Static("Index Season Digits")
|
||||||
|
yield Input(type="integer", id="index_season_digits_input", classes="four")
|
||||||
|
|
||||||
|
#7
|
||||||
|
yield Static("Index Episode Digits")
|
||||||
|
yield Input(type="integer", id="index_episode_digits_input", classes="four")
|
||||||
|
|
||||||
|
#8
|
||||||
|
yield Static("Indicator Season Digits")
|
||||||
|
yield Input(type="integer", id="indicator_season_digits_input", classes="four")
|
||||||
|
|
||||||
|
#9
|
||||||
|
yield Static("Indicator Edisode Digits")
|
||||||
|
yield Input(type="integer", id="indicator_episode_digits_input", classes="four")
|
||||||
|
|
||||||
|
# 10
|
||||||
|
yield Static(" ", classes="five")
|
||||||
|
|
||||||
|
# 11
|
||||||
|
yield Static("Shifted seasons", classes="two")
|
||||||
|
|
||||||
|
if self.__showDescriptor is not None:
|
||||||
|
yield Button("Add", id="button_add_shifted_season")
|
||||||
|
yield Button("Edit", id="button_edit_shifted_season")
|
||||||
|
yield Button("Delete", id="button_delete_shifted_season")
|
||||||
|
else:
|
||||||
|
yield Static(" ")
|
||||||
|
yield Static(" ")
|
||||||
|
yield Static(" ")
|
||||||
|
|
||||||
|
# 12
|
||||||
|
yield self.shiftedSeasonsTable
|
||||||
|
|
||||||
|
# 13
|
||||||
|
yield Static("File patterns", classes="five")
|
||||||
|
# 14
|
||||||
|
yield self.patternTable
|
||||||
|
|
||||||
|
# 15
|
||||||
|
yield Static(" ", classes="five")
|
||||||
|
|
||||||
|
# 16
|
||||||
|
yield Button("Save", id="save_button")
|
||||||
|
yield Button("Cancel", id="cancel_button")
|
||||||
|
|
||||||
|
|
||||||
|
yield Footer()
|
||||||
|
|
||||||
|
|
||||||
|
def getShowDescriptorFromInput(self) -> ShowDescriptor:
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
|
||||||
|
try:
|
||||||
|
if self.__showDescriptor:
|
||||||
|
kwargs[ShowDescriptor.ID_KEY] = int(self.__showDescriptor.getId())
|
||||||
|
else:
|
||||||
|
kwargs[ShowDescriptor.ID_KEY] = int(self.query_one("#id_input", Input).value)
|
||||||
|
except ValueError:
|
||||||
|
return None
|
||||||
|
|
||||||
|
try:
|
||||||
|
kwargs[ShowDescriptor.NAME_KEY] = str(self.query_one("#name_input", Input).value)
|
||||||
|
except ValueError:
|
||||||
|
pass
|
||||||
|
try:
|
||||||
|
kwargs[ShowDescriptor.YEAR_KEY] = int(self.query_one("#year_input", Input).value)
|
||||||
|
except ValueError:
|
||||||
|
pass
|
||||||
|
|
||||||
|
try:
|
||||||
|
kwargs[ShowDescriptor.INDEX_SEASON_DIGITS_KEY] = int(self.query_one("#index_season_digits_input", Input).value)
|
||||||
|
except ValueError:
|
||||||
|
pass
|
||||||
|
|
||||||
|
try:
|
||||||
|
kwargs[ShowDescriptor.INDEX_EPISODE_DIGITS_KEY] = int(self.query_one("#index_episode_digits_input", Input).value)
|
||||||
|
except ValueError:
|
||||||
|
pass
|
||||||
|
try:
|
||||||
|
kwargs[ShowDescriptor.INDICATOR_SEASON_DIGITS_KEY] = int(self.query_one("#indicator_season_digits_input", Input).value)
|
||||||
|
except ValueError:
|
||||||
|
pass
|
||||||
|
try:
|
||||||
|
kwargs[ShowDescriptor.INDICATOR_EPISODE_DIGITS_KEY] = int(self.query_one("#indicator_episode_digits_input", Input).value)
|
||||||
|
except ValueError:
|
||||||
|
pass
|
||||||
|
|
||||||
|
return ShowDescriptor(**kwargs)
|
||||||
|
|
||||||
|
|
||||||
|
# Event handler for button press
|
||||||
|
def on_button_pressed(self, event: Button.Pressed) -> None:
|
||||||
|
# Check if the button pressed is the one we are interested in
|
||||||
|
if event.button.id == "save_button":
|
||||||
|
|
||||||
|
showDescriptor = self.getShowDescriptorFromInput()
|
||||||
|
|
||||||
|
if not showDescriptor is None:
|
||||||
|
if self.__sc.updateShow(showDescriptor):
|
||||||
|
self.dismiss(showDescriptor)
|
||||||
|
else:
|
||||||
|
#TODO: Meldung
|
||||||
|
self.app.pop_screen()
|
||||||
|
|
||||||
|
if event.button.id == "cancel_button":
|
||||||
|
self.app.pop_screen()
|
||||||
|
|
||||||
|
|
||||||
|
if event.button.id == "identify_button":
|
||||||
|
|
||||||
|
showDescriptor = self.getShowDescriptorFromInput()
|
||||||
|
if not showDescriptor is None:
|
||||||
|
showName, showYear = self.__tc.getShowNameAndYear(showDescriptor.getId())
|
||||||
|
|
||||||
|
self.query_one("#name_input", Input).value = filterFilename(showName)
|
||||||
|
self.query_one("#year_input", Input).value = str(showYear)
|
||||||
|
|
||||||
|
|
||||||
|
if event.button.id == "button_add_shifted_season":
|
||||||
|
if not self.__showDescriptor is None:
|
||||||
|
self.app.push_screen(ShiftedSeasonDetailsScreen(showId = self.__showDescriptor.getId()), self.handle_update_shifted_season)
|
||||||
|
|
||||||
|
if event.button.id == "button_edit_shifted_season":
|
||||||
|
selectedShiftedSeasonObj = self.getSelectedShiftedSeasonObjFromInput()
|
||||||
|
if 'id' in selectedShiftedSeasonObj.keys():
|
||||||
|
self.app.push_screen(ShiftedSeasonDetailsScreen(showId = self.__showDescriptor.getId(), shiftedSeasonId=selectedShiftedSeasonObj['id']), self.handle_update_shifted_season)
|
||||||
|
|
||||||
|
if event.button.id == "button_delete_shifted_season":
|
||||||
|
selectedShiftedSeasonObj = self.getSelectedShiftedSeasonObjFromInput()
|
||||||
|
if 'id' in selectedShiftedSeasonObj.keys():
|
||||||
|
self.app.push_screen(ShiftedSeasonDeleteScreen(showId = self.__showDescriptor.getId(), shiftedSeasonId=selectedShiftedSeasonObj['id']), self.handle_delete_shifted_season)
|
||||||
|
|
||||||
|
|
||||||
|
def handle_update_shifted_season(self, screenResult):
|
||||||
|
self.updateShiftedSeasons()
|
||||||
|
|
||||||
|
def handle_delete_shifted_season(self, screenResult):
|
||||||
|
self.updateShiftedSeasons()
|
||||||
165
src/ffx/shows_screen.py
Normal file
165
src/ffx/shows_screen.py
Normal file
@@ -0,0 +1,165 @@
|
|||||||
|
from textual.screen import Screen
|
||||||
|
from textual.widgets import Header, Footer, Static, DataTable
|
||||||
|
from textual.containers import Grid
|
||||||
|
|
||||||
|
from .show_controller import ShowController
|
||||||
|
|
||||||
|
from .show_details_screen import ShowDetailsScreen
|
||||||
|
from .show_delete_screen import ShowDeleteScreen
|
||||||
|
|
||||||
|
from ffx.show_descriptor import ShowDescriptor
|
||||||
|
|
||||||
|
from textual.widgets._data_table import CellDoesNotExist
|
||||||
|
|
||||||
|
|
||||||
|
class ShowsScreen(Screen):
|
||||||
|
|
||||||
|
CSS = """
|
||||||
|
|
||||||
|
Grid {
|
||||||
|
grid-size: 1;
|
||||||
|
grid-rows: 2 auto;
|
||||||
|
height: 100%;
|
||||||
|
width: 100%;
|
||||||
|
padding: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable .datatable--cursor {
|
||||||
|
background: darkorange;
|
||||||
|
color: black;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable .datatable--header {
|
||||||
|
background: steelblue;
|
||||||
|
color: white;
|
||||||
|
}
|
||||||
|
|
||||||
|
#top {
|
||||||
|
height: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
#two {
|
||||||
|
column-span: 2;
|
||||||
|
row-span: 2;
|
||||||
|
tint: magenta 40%;
|
||||||
|
}
|
||||||
|
|
||||||
|
.box {
|
||||||
|
height: 100%;
|
||||||
|
border: solid green;
|
||||||
|
}
|
||||||
|
"""
|
||||||
|
|
||||||
|
BINDINGS = [
|
||||||
|
("e", "edit_show", "Edit Show"),
|
||||||
|
("n", "new_show", "New Show"),
|
||||||
|
("d", "delete_show", "Delete Show"),
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
super().__init__()
|
||||||
|
|
||||||
|
self.context = self.app.getContext()
|
||||||
|
self.Session = self.context['database']['session'] # convenience
|
||||||
|
|
||||||
|
self.__sc = ShowController(context = self.context)
|
||||||
|
|
||||||
|
|
||||||
|
def getSelectedShowId(self):
|
||||||
|
|
||||||
|
try:
|
||||||
|
# Fetch the currently selected row when 'Enter' is pressed
|
||||||
|
#selected_row_index = self.table.cursor_row
|
||||||
|
row_key, col_key = self.table.coordinate_to_cell_key(self.table.cursor_coordinate)
|
||||||
|
|
||||||
|
if row_key is not None:
|
||||||
|
selected_row_data = self.table.get_row(row_key)
|
||||||
|
|
||||||
|
return selected_row_data[0]
|
||||||
|
|
||||||
|
except CellDoesNotExist:
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def action_new_show(self):
|
||||||
|
self.app.push_screen(ShowDetailsScreen(), self.handle_new_screen)
|
||||||
|
|
||||||
|
def handle_new_screen(self, screenResult):
|
||||||
|
|
||||||
|
show = (screenResult['id'], screenResult['name'], screenResult['year'])
|
||||||
|
self.table.add_row(*map(str, show))
|
||||||
|
|
||||||
|
|
||||||
|
def action_edit_show(self):
|
||||||
|
|
||||||
|
selectedShowId = self.getSelectedShowId()
|
||||||
|
|
||||||
|
if selectedShowId is not None:
|
||||||
|
self.app.push_screen(ShowDetailsScreen(showId = selectedShowId), self.handle_edit_screen)
|
||||||
|
|
||||||
|
|
||||||
|
def handle_edit_screen(self, showDescriptor: ShowDescriptor):
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
row_key, col_key = self.table.coordinate_to_cell_key(self.table.cursor_coordinate)
|
||||||
|
|
||||||
|
self.table.update_cell(row_key, self.column_key_name, showDescriptor.getName())
|
||||||
|
self.table.update_cell(row_key, self.column_key_year, showDescriptor.getYear())
|
||||||
|
|
||||||
|
except CellDoesNotExist:
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def action_delete_show(self):
|
||||||
|
|
||||||
|
selectedShowId = self.getSelectedShowId()
|
||||||
|
|
||||||
|
if selectedShowId is not None:
|
||||||
|
self.app.push_screen(ShowDeleteScreen(showId = selectedShowId), self.handle_delete_show)
|
||||||
|
|
||||||
|
|
||||||
|
def handle_delete_show(self, showDescriptor: ShowDescriptor):
|
||||||
|
|
||||||
|
try:
|
||||||
|
row_key, col_key = self.table.coordinate_to_cell_key(self.table.cursor_coordinate)
|
||||||
|
self.table.remove_row(row_key)
|
||||||
|
|
||||||
|
except CellDoesNotExist:
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
def on_mount(self) -> None:
|
||||||
|
for show in self.__sc.getAllShows():
|
||||||
|
row = (int(show.id), show.name, show.year) # Convert each element to a string before adding
|
||||||
|
self.table.add_row(*map(str, row))
|
||||||
|
|
||||||
|
|
||||||
|
def compose(self):
|
||||||
|
|
||||||
|
# Create the DataTable widget
|
||||||
|
self.table = DataTable()
|
||||||
|
|
||||||
|
# Define the columns with headers
|
||||||
|
self.column_key_id = self.table.add_column("ID", width=10)
|
||||||
|
self.column_key_name = self.table.add_column("Name", width=50)
|
||||||
|
self.column_key_year = self.table.add_column("Year", width=10)
|
||||||
|
|
||||||
|
self.table.cursor_type = 'row'
|
||||||
|
|
||||||
|
yield Header()
|
||||||
|
|
||||||
|
with Grid():
|
||||||
|
|
||||||
|
yield Static("Shows")
|
||||||
|
|
||||||
|
yield self.table
|
||||||
|
|
||||||
|
yield Footer()
|
||||||
220
src/ffx/tag_controller.py
Normal file
220
src/ffx/tag_controller.py
Normal file
@@ -0,0 +1,220 @@
|
|||||||
|
import click
|
||||||
|
|
||||||
|
from ffx.model.track import Track
|
||||||
|
|
||||||
|
from ffx.model.media_tag import MediaTag
|
||||||
|
from ffx.model.track_tag import TrackTag
|
||||||
|
|
||||||
|
|
||||||
|
class TagController():
|
||||||
|
|
||||||
|
def __init__(self, context):
|
||||||
|
|
||||||
|
self.context = context
|
||||||
|
self.Session = self.context['database']['session'] # convenience
|
||||||
|
|
||||||
|
|
||||||
|
def updateMediaTag(self, patternId, tagKey, tagValue):
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
|
||||||
|
q = s.query(MediaTag).filter(MediaTag.pattern_id == int(patternId),
|
||||||
|
MediaTag.key == str(tagKey))
|
||||||
|
tag = q.first()
|
||||||
|
if tag:
|
||||||
|
tag.value = str(tagValue)
|
||||||
|
else:
|
||||||
|
tag = MediaTag(pattern_id = int(patternId),
|
||||||
|
key = str(tagKey),
|
||||||
|
value = str(tagValue))
|
||||||
|
s.add(tag)
|
||||||
|
s.commit()
|
||||||
|
|
||||||
|
return int(tag.id)
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"TagController.updateTrackTag(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
def updateTrackTag(self, trackId, tagKey, tagValue):
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
|
||||||
|
q = s.query(TrackTag).filter(TrackTag.track_id == int(trackId),
|
||||||
|
TrackTag.key == str(tagKey))
|
||||||
|
tag = q.first()
|
||||||
|
if tag:
|
||||||
|
tag.value = str(tagValue)
|
||||||
|
else:
|
||||||
|
tag = TrackTag(track_id = int(trackId),
|
||||||
|
key = str(tagKey),
|
||||||
|
value = str(tagValue))
|
||||||
|
s.add(tag)
|
||||||
|
s.commit()
|
||||||
|
|
||||||
|
return int(tag.id)
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"TagController.updateTrackTag(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
def deleteMediaTagByKey(self, patternId, tagKey):
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
|
||||||
|
q = s.query(MediaTag).filter(MediaTag.pattern_id == int(patternId),
|
||||||
|
MediaTag.key == str(tagKey))
|
||||||
|
if q.count():
|
||||||
|
tag = q.first()
|
||||||
|
s.delete(tag)
|
||||||
|
s.commit()
|
||||||
|
return True
|
||||||
|
else:
|
||||||
|
return False
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"TagController.deleteMediaTagByKey(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
def deleteTrackTagByKey(self, trackId, tagKey):
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
|
||||||
|
q = s.query(TrackTag).filter(TrackTag.track_id == int(trackId),
|
||||||
|
TrackTag.key == str(tagKey))
|
||||||
|
tag = q.first()
|
||||||
|
if tag:
|
||||||
|
s.delete(tag)
|
||||||
|
s.commit()
|
||||||
|
return True
|
||||||
|
else:
|
||||||
|
return False
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"TagController.deleteTrackTagByKey(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
def findAllMediaTags(self, patternId) -> dict:
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
|
||||||
|
q = s.query(MediaTag).filter(MediaTag.pattern_id == int(patternId))
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
return {t.key:t.value for t in q.all()}
|
||||||
|
else:
|
||||||
|
return {}
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"TagController.findAllMediaTags(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def findAllTrackTags(self, trackId) -> dict:
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
|
||||||
|
q = s.query(TrackTag).filter(TrackTag.track_id == int(trackId))
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
return {t.key:t.value for t in q.all()}
|
||||||
|
else:
|
||||||
|
return {}
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"TagController.findAllTracks(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def findMediaTag(self, trackId : int, trackKey : str) -> MediaTag:
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(Track).filter(MediaTag.track_id == int(trackId), MediaTag.key == str(trackKey))
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
return q.first()
|
||||||
|
else:
|
||||||
|
return None
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"TagController.findMediaTag(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
def findTrackTag(self, trackId : int, tagKey : str) -> TrackTag:
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(TrackTag).filter(TrackTag.track_id == int(trackId), TrackTag.key == str(tagKey))
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
return q.first()
|
||||||
|
else:
|
||||||
|
return None
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"TagController.findTrackTag(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def deleteMediaTag(self, tagId) -> bool:
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(MediaTag).filter(MediaTag.id == int(tagId))
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
|
||||||
|
tag = q.first()
|
||||||
|
|
||||||
|
s.delete(tag)
|
||||||
|
|
||||||
|
s.commit()
|
||||||
|
return True
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"TagController.deleteMediaTag(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
|
|
||||||
|
|
||||||
|
def deleteTrackTag(self, tagId : int) -> bool:
|
||||||
|
|
||||||
|
if type(tagId) is not int:
|
||||||
|
raise TypeError('TagController.deleteTrackTag(): Argument tagId is required to be of type int')
|
||||||
|
|
||||||
|
try:
|
||||||
|
s = self.Session()
|
||||||
|
q = s.query(TrackTag).filter(TrackTag.id == int(tagId))
|
||||||
|
|
||||||
|
if q.count():
|
||||||
|
|
||||||
|
tag = q.first()
|
||||||
|
|
||||||
|
s.delete(tag)
|
||||||
|
|
||||||
|
s.commit()
|
||||||
|
return True
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
|
except Exception as ex:
|
||||||
|
raise click.ClickException(f"TagController.deleteTrackTag(): {repr(ex)}")
|
||||||
|
finally:
|
||||||
|
s.close()
|
||||||
98
src/ffx/tag_delete_screen.py
Normal file
98
src/ffx/tag_delete_screen.py
Normal file
@@ -0,0 +1,98 @@
|
|||||||
|
from textual.screen import Screen
|
||||||
|
from textual.widgets import Header, Footer, Static, Button
|
||||||
|
from textual.containers import Grid
|
||||||
|
|
||||||
|
|
||||||
|
# Screen[dict[int, str, int]]
|
||||||
|
class TagDeleteScreen(Screen):
|
||||||
|
|
||||||
|
CSS = """
|
||||||
|
|
||||||
|
Grid {
|
||||||
|
grid-size: 4 9;
|
||||||
|
grid-rows: 2 2 2 2 2 2 2 2 2;
|
||||||
|
grid-columns: 30 30 30 30;
|
||||||
|
height: 100%;
|
||||||
|
width: 100%;
|
||||||
|
padding: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
Input {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
Button {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
#toplabel {
|
||||||
|
height: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
.two {
|
||||||
|
column-span: 2;
|
||||||
|
}
|
||||||
|
.three {
|
||||||
|
column-span: 3;
|
||||||
|
}
|
||||||
|
.four {
|
||||||
|
column-span: 4;
|
||||||
|
}
|
||||||
|
.five {
|
||||||
|
column-span: 5;
|
||||||
|
}
|
||||||
|
|
||||||
|
.box {
|
||||||
|
height: 100%;
|
||||||
|
border: solid green;
|
||||||
|
}
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, key=None, value=None):
|
||||||
|
super().__init__()
|
||||||
|
self.__key = key
|
||||||
|
self.__value = value
|
||||||
|
|
||||||
|
|
||||||
|
def on_mount(self):
|
||||||
|
|
||||||
|
self.query_one("#keylabel", Static).update(str(self.__key))
|
||||||
|
self.query_one("#valuelabel", Static).update(str(self.__value))
|
||||||
|
|
||||||
|
|
||||||
|
def compose(self):
|
||||||
|
|
||||||
|
yield Header()
|
||||||
|
|
||||||
|
with Grid():
|
||||||
|
|
||||||
|
#1
|
||||||
|
yield Static(f"Are you sure to delete this tag ?", id="toplabel", classes="five")
|
||||||
|
|
||||||
|
#2
|
||||||
|
yield Static("Key")
|
||||||
|
yield Static(" ", id="keylabel", classes="four")
|
||||||
|
|
||||||
|
#3
|
||||||
|
yield Static("Value")
|
||||||
|
yield Static(" ", id="valuelabel", classes="four")
|
||||||
|
|
||||||
|
#4
|
||||||
|
yield Static(" ", classes="five")
|
||||||
|
|
||||||
|
#9
|
||||||
|
yield Button("Delete", id="delete_button")
|
||||||
|
yield Button("Cancel", id="cancel_button")
|
||||||
|
|
||||||
|
yield Footer()
|
||||||
|
|
||||||
|
|
||||||
|
# Event handler for button press
|
||||||
|
def on_button_pressed(self, event: Button.Pressed) -> None:
|
||||||
|
|
||||||
|
if event.button.id == "delete_button":
|
||||||
|
|
||||||
|
tag = (self.__key, self.__value)
|
||||||
|
self.dismiss(tag)
|
||||||
|
|
||||||
|
if event.button.id == "cancel_button":
|
||||||
|
self.app.pop_screen()
|
||||||
|
|
||||||
132
src/ffx/tag_details_screen.py
Normal file
132
src/ffx/tag_details_screen.py
Normal file
@@ -0,0 +1,132 @@
|
|||||||
|
from textual.screen import Screen
|
||||||
|
from textual.widgets import Header, Footer, Static, Button, Input
|
||||||
|
from textual.containers import Grid
|
||||||
|
|
||||||
|
|
||||||
|
# Screen[dict[int, str, int]]
|
||||||
|
class TagDetailsScreen(Screen):
|
||||||
|
|
||||||
|
CSS = """
|
||||||
|
|
||||||
|
Grid {
|
||||||
|
grid-size: 5 20;
|
||||||
|
grid-rows: 2 2 2 2 2 3 2 2 2 2 2 6 2 2 6 2 2 2 2 6;
|
||||||
|
grid-columns: 25 25 25 25 225;
|
||||||
|
height: 100%;
|
||||||
|
width: 100%;
|
||||||
|
padding: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
Input {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
Button {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
SelectionList {
|
||||||
|
border: none;
|
||||||
|
min-height: 6;
|
||||||
|
}
|
||||||
|
Select {
|
||||||
|
border: none;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable {
|
||||||
|
min-height: 6;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable .datatable--cursor {
|
||||||
|
background: darkorange;
|
||||||
|
color: black;
|
||||||
|
}
|
||||||
|
|
||||||
|
DataTable .datatable--header {
|
||||||
|
background: steelblue;
|
||||||
|
color: white;
|
||||||
|
}
|
||||||
|
|
||||||
|
#toplabel {
|
||||||
|
height: 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
.two {
|
||||||
|
column-span: 2;
|
||||||
|
}
|
||||||
|
.three {
|
||||||
|
column-span: 3;
|
||||||
|
}
|
||||||
|
|
||||||
|
.four {
|
||||||
|
column-span: 4;
|
||||||
|
}
|
||||||
|
.five {
|
||||||
|
column-span: 5;
|
||||||
|
}
|
||||||
|
|
||||||
|
.box {
|
||||||
|
height: 100%;
|
||||||
|
border: solid green;
|
||||||
|
}
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, key=None, value=None):
|
||||||
|
super().__init__()
|
||||||
|
self.__key = key
|
||||||
|
self.__value = value
|
||||||
|
|
||||||
|
|
||||||
|
def on_mount(self):
|
||||||
|
|
||||||
|
if self.__key is not None:
|
||||||
|
self.query_one("#key_input", Input).value = str(self.__key)
|
||||||
|
|
||||||
|
if self.__value is not None:
|
||||||
|
self.query_one("#value_input", Input).value = str(self.__value)
|
||||||
|
|
||||||
|
|
||||||
|
def compose(self):
|
||||||
|
|
||||||
|
yield Header()
|
||||||
|
|
||||||
|
with Grid():
|
||||||
|
|
||||||
|
# 8
|
||||||
|
yield Static("Key")
|
||||||
|
yield Input(id="key_input", classes="four")
|
||||||
|
|
||||||
|
yield Static("Value")
|
||||||
|
yield Input(id="value_input", classes="four")
|
||||||
|
|
||||||
|
# 17
|
||||||
|
yield Static(" ", classes="five")
|
||||||
|
|
||||||
|
# 18
|
||||||
|
yield Button("Save", id="save_button")
|
||||||
|
yield Button("Cancel", id="cancel_button")
|
||||||
|
|
||||||
|
# 19
|
||||||
|
yield Static(" ", classes="five")
|
||||||
|
|
||||||
|
# 20
|
||||||
|
yield Static(" ", classes="five", id="messagestatic")
|
||||||
|
|
||||||
|
yield Footer(id="footer")
|
||||||
|
|
||||||
|
|
||||||
|
def getTagFromInput(self):
|
||||||
|
|
||||||
|
tagKey = self.query_one("#key_input", Input).value
|
||||||
|
tagValue = self.query_one("#value_input", Input).value
|
||||||
|
|
||||||
|
return (tagKey, tagValue)
|
||||||
|
|
||||||
|
|
||||||
|
# Event handler for button press
|
||||||
|
def on_button_pressed(self, event: Button.Pressed) -> None:
|
||||||
|
|
||||||
|
# Check if the button pressed is the one we are interested in
|
||||||
|
if event.button.id == "save_button":
|
||||||
|
self.dismiss(self.getTagFromInput())
|
||||||
|
|
||||||
|
if event.button.id == "cancel_button":
|
||||||
|
self.app.pop_screen()
|
||||||
64
src/ffx/test/_basename_combinator_1.py
Normal file
64
src/ffx/test/_basename_combinator_1.py
Normal file
@@ -0,0 +1,64 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .basename_combinator import BasenameCombinator
|
||||||
|
|
||||||
|
from .indicator_combinator import IndicatorCombinator
|
||||||
|
from .label_combinator import LabelCombinator
|
||||||
|
|
||||||
|
class BasenameCombinator2(BasenameCombinator):
|
||||||
|
"""documentation_site"""
|
||||||
|
|
||||||
|
VARIANT = 'B2'
|
||||||
|
|
||||||
|
# def __init__(self, SubCombinators: dict = {}, context = None):
|
||||||
|
def __init__(self, context = None):
|
||||||
|
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return BasenameCombinator2.VARIANT
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
return ''
|
||||||
|
|
||||||
|
def assertFunc(self, mediaDescriptor: MediaDescriptor):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
|
||||||
|
for L in LabelCombinator.getAllClassReferences():
|
||||||
|
# for I in IndicatorCombinator.getAllClassReferences():
|
||||||
|
# for S in SiteCombinator.getAllClassReferences():
|
||||||
|
# for T in TitleCombinator.getAllClassReferences():
|
||||||
|
#
|
||||||
|
|
||||||
|
l = L(self._context)
|
||||||
|
|
||||||
|
yieldObj = {}
|
||||||
|
|
||||||
|
yieldObj['identifier'] = self.getIdentifier()
|
||||||
|
|
||||||
|
yieldObj['variants'] = [self.getVariant(),
|
||||||
|
l.getVariant()]
|
||||||
|
|
||||||
|
yieldObj['payload'] = {'label': l.getPayload()}
|
||||||
|
|
||||||
|
yieldObj['assertSelectors'] = ['B', 'L']
|
||||||
|
|
||||||
|
yieldObj['assertFuncs'] = [self.assertFunc,
|
||||||
|
l.assertFunc]
|
||||||
|
|
||||||
|
yieldObj['shouldFail'] = (self.shouldFail()
|
||||||
|
| l.shouldFail())
|
||||||
|
|
||||||
|
yield yieldObj
|
||||||
35
src/ffx/test/basename_combinator.py
Normal file
35
src/ffx/test/basename_combinator.py
Normal file
@@ -0,0 +1,35 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
class BasenameCombinator():
|
||||||
|
|
||||||
|
IDENTIFIER = 'basename'
|
||||||
|
|
||||||
|
BASENAME = 'media'
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getIdentifier(self):
|
||||||
|
return BasenameCombinator.IDENTIFIER
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def list():
|
||||||
|
basePath = os.path.dirname(__file__)
|
||||||
|
return [os.path.basename(p)[20:-3]
|
||||||
|
for p
|
||||||
|
in glob.glob(f"{ basePath }/basename_combinator_*.py", recursive = True)
|
||||||
|
if p != __file__]
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getClassReference(identifier):
|
||||||
|
importlib.import_module(f"ffx.test.basename_combinator_{ identifier }")
|
||||||
|
for name, obj in inspect.getmembers(sys.modules[f"ffx.test.basename_combinator_{ identifier }"]):
|
||||||
|
#HINT: Excluding MediaCombinator as it seems to be included by import (?)
|
||||||
|
if inspect.isclass(obj) and name != 'BasenameCombinator' and name.startswith('BasenameCombinator'):
|
||||||
|
return obj
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getAllClassReferences():
|
||||||
|
return [BasenameCombinator.getClassReference(i) for i in BasenameCombinator.list()]
|
||||||
107
src/ffx/test/basename_combinator_0.py
Normal file
107
src/ffx/test/basename_combinator_0.py
Normal file
@@ -0,0 +1,107 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .basename_combinator import BasenameCombinator
|
||||||
|
|
||||||
|
from .indicator_combinator import IndicatorCombinator
|
||||||
|
from .label_combinator import LabelCombinator
|
||||||
|
|
||||||
|
|
||||||
|
class BasenameCombinator0(BasenameCombinator):
|
||||||
|
"""base[_indicator]"""
|
||||||
|
|
||||||
|
VARIANT = 'B0'
|
||||||
|
|
||||||
|
# def __init__(self, SubCombinators: dict = {}, context = None):
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return BasenameCombinator0.VARIANT
|
||||||
|
|
||||||
|
def getPayload(self, indicator = '', label = ''):
|
||||||
|
|
||||||
|
basename = BasenameCombinator.BASENAME
|
||||||
|
expectedBasename = label if label else BasenameCombinator.BASENAME
|
||||||
|
|
||||||
|
if indicator:
|
||||||
|
basename += f"_{indicator}"
|
||||||
|
expectedBasename += f"_{indicator}"
|
||||||
|
|
||||||
|
return {'basename': basename,
|
||||||
|
'label': label,
|
||||||
|
'expectedBasename': expectedBasename}
|
||||||
|
|
||||||
|
|
||||||
|
def assertFunc(self, indicator = '', label = ''):
|
||||||
|
|
||||||
|
def f(testObj: dict = {}):
|
||||||
|
|
||||||
|
if not 'filenames' in testObj.keys():
|
||||||
|
raise KeyError("testObj does not contain key 'filenames'")
|
||||||
|
|
||||||
|
fNames = testObj['filenames']
|
||||||
|
|
||||||
|
assert len(fNames) == 1, "More than one result file was created"
|
||||||
|
|
||||||
|
resultFilename = fNames[0]
|
||||||
|
|
||||||
|
fTokens = resultFilename.split('.')
|
||||||
|
|
||||||
|
resultBasename = '.'.join(fTokens[:-1])
|
||||||
|
resultExtension = fTokens[-1]
|
||||||
|
|
||||||
|
if not indicator and not label:
|
||||||
|
|
||||||
|
assert resultBasename == BasenameCombinator.BASENAME, f"Result basename is not {BasenameCombinator.BASENAME}"
|
||||||
|
if not indicator and label:
|
||||||
|
assert resultBasename == label, f"Result basename is not {label}"
|
||||||
|
if indicator and not label:
|
||||||
|
assert resultBasename == f"{BasenameCombinator.BASENAME}_{indicator}", f"Result basename is not {BasenameCombinator.BASENAME}_{indicator}"
|
||||||
|
if indicator and label:
|
||||||
|
assert resultBasename == f"{label}_{indicator}", f"Result basename is not {label}_{indicator}"
|
||||||
|
|
||||||
|
return f
|
||||||
|
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
|
||||||
|
ic = IndicatorCombinator(self._context)
|
||||||
|
|
||||||
|
for L in LabelCombinator.getAllClassReferences():
|
||||||
|
for i in ic.getYield():
|
||||||
|
|
||||||
|
l = L(self._context)
|
||||||
|
|
||||||
|
indicator = i['indicator']
|
||||||
|
indicatorVariant = i['variant']
|
||||||
|
|
||||||
|
yieldObj = {}
|
||||||
|
|
||||||
|
yieldObj['identifier'] = self.getIdentifier()
|
||||||
|
|
||||||
|
yieldObj['variants'] = [self.getVariant(),
|
||||||
|
l.getVariant(),
|
||||||
|
indicatorVariant]
|
||||||
|
|
||||||
|
yieldObj['payload'] = self.getPayload(indicator = indicator,
|
||||||
|
label = l.getPayload())
|
||||||
|
|
||||||
|
yieldObj['assertSelectors'] = ['B', 'L', 'I']
|
||||||
|
|
||||||
|
yieldObj['assertFuncs'] = [self.assertFunc(indicator, l.getPayload()), l.assertFunc, ic.assertFunc]
|
||||||
|
|
||||||
|
yieldObj['shouldFail'] = (self.shouldFail()
|
||||||
|
| l.shouldFail()
|
||||||
|
| ic.shouldFail())
|
||||||
|
|
||||||
|
yield yieldObj
|
||||||
159
src/ffx/test/basename_combinator_2.py
Normal file
159
src/ffx/test/basename_combinator_2.py
Normal file
@@ -0,0 +1,159 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .basename_combinator import BasenameCombinator
|
||||||
|
|
||||||
|
from .indicator_combinator import IndicatorCombinator
|
||||||
|
from .label_combinator import LabelCombinator
|
||||||
|
from .title_combinator import TitleCombinator
|
||||||
|
from .release_combinator import ReleaseCombinator
|
||||||
|
from .show_combinator import ShowCombinator
|
||||||
|
|
||||||
|
|
||||||
|
class BasenameCombinator2(BasenameCombinator):
|
||||||
|
"""show[_indicator]_group"""
|
||||||
|
|
||||||
|
VARIANT = 'B2'
|
||||||
|
|
||||||
|
# def __init__(self, SubCombinators: dict = {}, context = None):
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return BasenameCombinator2.VARIANT
|
||||||
|
|
||||||
|
#
|
||||||
|
# SHOW_LIST = [
|
||||||
|
# 'Boruto: Naruto Next Generations (2017)',
|
||||||
|
# 'The Rising of the Shield Hero (2019)',
|
||||||
|
# 'Scrubs - Die Anfänger (2001)'
|
||||||
|
# ]
|
||||||
|
#
|
||||||
|
# RELEASE_LIST = [
|
||||||
|
# ".GerEngSub.AAC.1080pINDICATOR.WebDL.x264-Tanuki",
|
||||||
|
# ".German.AC3.DL.1080pINDICATOR.BluRay.x264-AST4u",
|
||||||
|
# "-720pINDICATOR"
|
||||||
|
# ]
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self,
|
||||||
|
indicator = '',
|
||||||
|
label = '',
|
||||||
|
show = '',
|
||||||
|
release = ''):
|
||||||
|
|
||||||
|
if label:
|
||||||
|
basename = label
|
||||||
|
expectedBasename = label
|
||||||
|
if indicator:
|
||||||
|
basename += f"_{indicator}"
|
||||||
|
expectedBasename += f"_{indicator}"
|
||||||
|
else:
|
||||||
|
basename = show+release
|
||||||
|
expectedBasename = basename
|
||||||
|
|
||||||
|
return {'basename': basename,
|
||||||
|
'label': label,
|
||||||
|
'expectedBasename': expectedBasename}
|
||||||
|
|
||||||
|
|
||||||
|
def createAssertFunc(self,
|
||||||
|
indicator = '',
|
||||||
|
label = '',
|
||||||
|
show = '',
|
||||||
|
release = ''):
|
||||||
|
|
||||||
|
def f(testObj: dict = {}):
|
||||||
|
|
||||||
|
if not 'filenames' in testObj.keys():
|
||||||
|
raise KeyError("testObj does not contain key 'filenames'")
|
||||||
|
|
||||||
|
fNames = testObj['filenames']
|
||||||
|
|
||||||
|
assert len(fNames) == 1, "More than one result file was created"
|
||||||
|
|
||||||
|
resultFilename = fNames[0]
|
||||||
|
|
||||||
|
fTokens = resultFilename.split('.')
|
||||||
|
|
||||||
|
resultBasename = '.'.join(fTokens[:-1])
|
||||||
|
resultExtension = fTokens[-1]
|
||||||
|
|
||||||
|
if not indicator and not label:
|
||||||
|
assert resultBasename == show+release, f"Result basename is not {show+release}"
|
||||||
|
elif not indicator and label:
|
||||||
|
assert resultBasename == label, f"Result basename is not {label}"
|
||||||
|
elif indicator and not label:
|
||||||
|
assert resultBasename == show+release, f"Result basename is not {show+release}"
|
||||||
|
elif indicator and label:
|
||||||
|
assert resultBasename == f"{label}_{indicator}", f"Result basename is not {label}_{indicator}"
|
||||||
|
|
||||||
|
return f
|
||||||
|
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
|
||||||
|
ic = IndicatorCombinator(self._context)
|
||||||
|
sc = ShowCombinator(self._context)
|
||||||
|
|
||||||
|
for L in LabelCombinator.getAllClassReferences():
|
||||||
|
for iy in ic.getYield():
|
||||||
|
|
||||||
|
indicator = iy['indicator']
|
||||||
|
indicatorVariant = iy['variant']
|
||||||
|
|
||||||
|
rc = ReleaseCombinator(self._context, indicator=indicator)
|
||||||
|
|
||||||
|
for sy in sc.getYield():
|
||||||
|
for ry in rc.getYield():
|
||||||
|
|
||||||
|
l = L(self._context)
|
||||||
|
|
||||||
|
show = sy['show']
|
||||||
|
showVariant = sy['variant']
|
||||||
|
|
||||||
|
release = ry['release']
|
||||||
|
releaseVariant = ry['variant']
|
||||||
|
|
||||||
|
yieldObj = {}
|
||||||
|
|
||||||
|
yieldObj['identifier'] = self.getIdentifier()
|
||||||
|
|
||||||
|
yieldObj['variants'] = [self.getVariant(),
|
||||||
|
l.getVariant(),
|
||||||
|
indicatorVariant,
|
||||||
|
showVariant,
|
||||||
|
releaseVariant]
|
||||||
|
|
||||||
|
yieldObj['payload'] = self.getPayload(indicator = indicator,
|
||||||
|
label = l.getPayload(),
|
||||||
|
show = show,
|
||||||
|
release = release)
|
||||||
|
|
||||||
|
yieldObj['assertSelectors'] = ['B', 'L', 'I', 'S', 'R']
|
||||||
|
|
||||||
|
yieldObj['assertFuncs'] = [self.createAssertFunc(indicator,
|
||||||
|
l.getPayload(),
|
||||||
|
show = show,
|
||||||
|
release = release),
|
||||||
|
l.assertFunc,
|
||||||
|
ic.assertFunc,
|
||||||
|
sc.assertFunc,
|
||||||
|
rc.assertFunc]
|
||||||
|
|
||||||
|
yieldObj['shouldFail'] = (self.shouldFail()
|
||||||
|
| l.shouldFail()
|
||||||
|
| ic.shouldFail()
|
||||||
|
| sc.shouldFail()
|
||||||
|
| rc.shouldFail())
|
||||||
|
|
||||||
|
yield yieldObj
|
||||||
13
src/ffx/test/combinator.py
Normal file
13
src/ffx/test/combinator.py
Normal file
@@ -0,0 +1,13 @@
|
|||||||
|
class Combinator():
|
||||||
|
|
||||||
|
def __init__(self, SubCombinations: dict):
|
||||||
|
self._SubCombinators = SubCombinations
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def assertFunc(self, testObj):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def getYield(yieldObj: dict):
|
||||||
|
pass
|
||||||
35
src/ffx/test/disposition_combinator_2.py
Normal file
35
src/ffx/test/disposition_combinator_2.py
Normal file
@@ -0,0 +1,35 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
class DispositionCombinator2():
|
||||||
|
|
||||||
|
IDENTIFIER = 'disposition2'
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getIdentifier(self):
|
||||||
|
return DispositionCombinator2.IDENTIFIER
|
||||||
|
def getVariant(self):
|
||||||
|
return self._variant
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def list():
|
||||||
|
basePath = os.path.dirname(__file__)
|
||||||
|
return [os.path.basename(p)[25:-3]
|
||||||
|
for p
|
||||||
|
in glob.glob(f"{ basePath }/disposition_combinator_2_*.py", recursive = True)
|
||||||
|
if p != __file__]
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getClassReference(identifier):
|
||||||
|
importlib.import_module(f"ffx.test.disposition_combinator_2_{ identifier }")
|
||||||
|
for name, obj in inspect.getmembers(sys.modules[f"ffx.test.disposition_combinator_2_{ identifier }"]):
|
||||||
|
#HINT: Excluding DispositionCombination as it seems to be included by import (?)
|
||||||
|
if inspect.isclass(obj) and name != 'DispositionCombinator2' and name.startswith('DispositionCombinator2'):
|
||||||
|
return obj
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getAllClassReferences():
|
||||||
|
return [DispositionCombinator2.getClassReference(i) for i in DispositionCombinator2.list()]
|
||||||
76
src/ffx/test/disposition_combinator_2_0.py
Normal file
76
src/ffx/test/disposition_combinator_2_0.py
Normal file
@@ -0,0 +1,76 @@
|
|||||||
|
import os, sys, importlib, glob, inspect
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from .disposition_combinator_2 import DispositionCombinator2
|
||||||
|
|
||||||
|
|
||||||
|
class DispositionCombinator20(DispositionCombinator2):
|
||||||
|
|
||||||
|
# COMMENT
|
||||||
|
# DESCRIPTIONS
|
||||||
|
|
||||||
|
VARIANT = 'D00'
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return DispositionCombinator20.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
|
||||||
|
subtrack0 = set()
|
||||||
|
subtrack1 = set()
|
||||||
|
|
||||||
|
#NOTE: Current ffmpeg version will not set most of the dispositions on arbitrary tracks
|
||||||
|
# so some checks for preserved dispositions are omitted for now
|
||||||
|
if self.__createPresets:
|
||||||
|
subtrack0.add(TrackDisposition.FORCED) # COMMENT
|
||||||
|
# subtrack1.add(TrackDisposition.DESCRIPTIONS) # DESCRIPTIONS
|
||||||
|
|
||||||
|
return (subtrack0,
|
||||||
|
subtrack1)
|
||||||
|
|
||||||
|
def createAssertFunc(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (not trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has set default disposition"
|
||||||
|
assert (trackDescriptors[0].getDispositionFlag(TrackDisposition.FORCED)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has not preserved set 'forced' disposition"
|
||||||
|
# source subIndex 1
|
||||||
|
assert (not trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has set default disposition"
|
||||||
|
# assert (trackDescriptors[1].getDispositionFlag(TrackDisposition.DESCRIPTIONS)
|
||||||
|
# ), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has not preserved set 'descriptions' disposition"
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (not trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has set default disposition"
|
||||||
|
# source subIndex 1
|
||||||
|
assert (not trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has set default disposition"
|
||||||
|
|
||||||
|
return f
|
||||||
|
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
79
src/ffx/test/disposition_combinator_2_1.py
Normal file
79
src/ffx/test/disposition_combinator_2_1.py
Normal file
@@ -0,0 +1,79 @@
|
|||||||
|
import os, sys, importlib, glob, inspect
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from .disposition_combinator_2 import DispositionCombinator2
|
||||||
|
|
||||||
|
|
||||||
|
class DispositionCombinator21(DispositionCombinator2):
|
||||||
|
|
||||||
|
VARIANT = 'D10'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return DispositionCombinator21.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
subtrack0 = set()
|
||||||
|
subtrack1 = set([TrackDisposition.DEFAULT])
|
||||||
|
else:
|
||||||
|
subtrack0 = set([TrackDisposition.DEFAULT])
|
||||||
|
subtrack1 = set()
|
||||||
|
|
||||||
|
#NOTE: Current ffmpeg version will not set most of the dispositions on arbitrary tracks
|
||||||
|
# so some checks for preserved dispositions are omitted for now
|
||||||
|
if self.__createPresets:
|
||||||
|
# subtrack0.add(TrackDisposition.COMMENT) # COMMENT
|
||||||
|
subtrack1.add(TrackDisposition.FORCED) # DESCRIPTIONS
|
||||||
|
|
||||||
|
return (subtrack0,
|
||||||
|
subtrack1)
|
||||||
|
|
||||||
|
def createAssertFunc(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has not set default disposition"
|
||||||
|
# assert (trackDescriptors[0].getDispositionFlag(TrackDisposition.COMMENT)
|
||||||
|
# ), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has not preserved set 'comment' disposition"
|
||||||
|
# source subIndex 1
|
||||||
|
assert not (trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has set default disposition"
|
||||||
|
assert (trackDescriptors[1].getDispositionFlag(TrackDisposition.FORCED)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has not preserved set 'forced' disposition"
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has not set default disposition"
|
||||||
|
# source subIndex 1
|
||||||
|
assert (not trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has set default disposition"
|
||||||
|
|
||||||
|
return f
|
||||||
|
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
79
src/ffx/test/disposition_combinator_2_2.py
Normal file
79
src/ffx/test/disposition_combinator_2_2.py
Normal file
@@ -0,0 +1,79 @@
|
|||||||
|
import os, sys, importlib, glob, inspect
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from .disposition_combinator_2 import DispositionCombinator2
|
||||||
|
|
||||||
|
|
||||||
|
class DispositionCombinator22(DispositionCombinator2):
|
||||||
|
|
||||||
|
VARIANT = 'D01'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return DispositionCombinator22.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
subtrack0 = set([TrackDisposition.DEFAULT])
|
||||||
|
subtrack1 = set()
|
||||||
|
else:
|
||||||
|
subtrack0 = set()
|
||||||
|
subtrack1 = set([TrackDisposition.DEFAULT])
|
||||||
|
|
||||||
|
#NOTE: Current ffmpeg version will not set most of the dispositions on arbitrary tracks
|
||||||
|
# so some checks for preserved dispositions are omitted for now
|
||||||
|
if self.__createPresets:
|
||||||
|
subtrack0.add(TrackDisposition.FORCED) # COMMENT
|
||||||
|
# subtrack1.add(TrackDisposition.DESCRIPTIONS) # DESCRIPTIONS
|
||||||
|
|
||||||
|
return (subtrack0,
|
||||||
|
subtrack1)
|
||||||
|
|
||||||
|
|
||||||
|
def createAssertFunc(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (not trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has set default disposition"
|
||||||
|
assert (trackDescriptors[0].getDispositionFlag(TrackDisposition.FORCED)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has not preserved set 'descriptions' disposition"
|
||||||
|
# source subIndex 1
|
||||||
|
assert (trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has not set default disposition"
|
||||||
|
# assert (trackDescriptors[1].getDispositionFlag(TrackDisposition.DESCRIPTIONS)
|
||||||
|
# ), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has not preserved 'forced' disposition"
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (not trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has set default disposition"
|
||||||
|
# source subIndex 1
|
||||||
|
assert (trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has not set default disposition"
|
||||||
|
|
||||||
|
return f
|
||||||
|
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
43
src/ffx/test/disposition_combinator_2_3 .py
Normal file
43
src/ffx/test/disposition_combinator_2_3 .py
Normal file
@@ -0,0 +1,43 @@
|
|||||||
|
import os, sys, importlib, glob, inspect
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from .disposition_combinator_2 import DispositionCombinator2
|
||||||
|
|
||||||
|
|
||||||
|
class DispositionCombinator23(DispositionCombinator2):
|
||||||
|
|
||||||
|
VARIANT = 'D11'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return DispositionCombinator23.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
|
||||||
|
subtrack0 = set([TrackDisposition.DEFAULT])
|
||||||
|
subtrack1 = set([TrackDisposition.DEFAULT])
|
||||||
|
|
||||||
|
#NOTE: Current ffmpeg version will not set most of the dispositions on arbitrary tracks
|
||||||
|
# so some checks for preserved dispositions are omitted for now
|
||||||
|
if self.__createPresets:
|
||||||
|
# subtrack0.add(TrackDisposition.COMMENT) # COMMENT
|
||||||
|
subtrack1.add(TrackDisposition.FORCED) # DESCRIPTIONS
|
||||||
|
|
||||||
|
return (subtrack0,
|
||||||
|
subtrack1)
|
||||||
|
|
||||||
|
#TODO: tmdb cases
|
||||||
|
def createAssertFunc(self):
|
||||||
|
def f(assertObj: dict = {}):
|
||||||
|
pass
|
||||||
|
return f
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return True
|
||||||
34
src/ffx/test/disposition_combinator_3.py
Normal file
34
src/ffx/test/disposition_combinator_3.py
Normal file
@@ -0,0 +1,34 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
class DispositionCombinator3():
|
||||||
|
|
||||||
|
IDENTIFIER = 'disposition3'
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getIdentifier(self):
|
||||||
|
return DispositionCombinator3.IDENTIFIER
|
||||||
|
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def list():
|
||||||
|
basePath = os.path.dirname(__file__)
|
||||||
|
return [os.path.basename(p)[25:-3]
|
||||||
|
for p
|
||||||
|
in glob.glob(f"{ basePath }/disposition_combinator_3_*.py", recursive = True)
|
||||||
|
if p != __file__]
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getClassReference(identifier):
|
||||||
|
importlib.import_module(f"ffx.test.disposition_combinator_3_{ identifier }")
|
||||||
|
for name, obj in inspect.getmembers(sys.modules[f"ffx.test.disposition_combinator_3_{ identifier }"]):
|
||||||
|
#HINT: Excluding DispositionCombination as it seems to be included by import (?)
|
||||||
|
if inspect.isclass(obj) and name != 'DispositionCombinator3' and name.startswith('DispositionCombinator3'):
|
||||||
|
return obj
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getAllClassReferences():
|
||||||
|
return [DispositionCombinator3.getClassReference(i) for i in DispositionCombinator3.list()]
|
||||||
92
src/ffx/test/disposition_combinator_3_0.py
Normal file
92
src/ffx/test/disposition_combinator_3_0.py
Normal file
@@ -0,0 +1,92 @@
|
|||||||
|
import os, sys, importlib, glob, inspect
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from .disposition_combinator_3 import DispositionCombinator3
|
||||||
|
|
||||||
|
|
||||||
|
class DispositionCombinator30(DispositionCombinator3):
|
||||||
|
|
||||||
|
VARIANT = 'D000'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return DispositionCombinator30.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
|
||||||
|
subtrack0 = set()
|
||||||
|
subtrack1 = set()
|
||||||
|
subtrack2 = set()
|
||||||
|
|
||||||
|
#NOTE: Current ffmpeg version will not set most of the dispositions on arbitrary tracks
|
||||||
|
# so some checks for preserved dispositions are omitted for now
|
||||||
|
if self.__createPresets:
|
||||||
|
subtrack0.add(TrackDisposition.FORCED) # COMMENT
|
||||||
|
# subtrack1.add(TrackDisposition.DESCRIPTIONS) # DESCRIPTIONS
|
||||||
|
# subtrack2.add(TrackDisposition.HEARING_IMPAIRED) # HEARING_IMPAIRED
|
||||||
|
|
||||||
|
return (subtrack0,
|
||||||
|
subtrack1,
|
||||||
|
subtrack2)
|
||||||
|
|
||||||
|
|
||||||
|
def createAssertFunc(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert not (trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has set default disposition"
|
||||||
|
assert (trackDescriptors[0].getDispositionFlag(TrackDisposition.FORCED)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has not preserved default disposition"
|
||||||
|
|
||||||
|
# source subIndex 1
|
||||||
|
assert not (trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has set default disposition"
|
||||||
|
# assert (trackDescriptors[1].getDispositionFlag(TrackDisposition.DESCRIPTIONS)
|
||||||
|
# ), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has not preserved default disposition"
|
||||||
|
|
||||||
|
# source subIndex 2
|
||||||
|
assert not (trackDescriptors[2].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #2 index={trackDescriptors[2].getIndex()} [{trackDescriptors[2].getType().label()}:{trackDescriptors[2].getSubIndex()}] has set default disposition"
|
||||||
|
# assert (trackDescriptors[2].getDispositionFlag(TrackDisposition.HEARING_IMPAIRED)
|
||||||
|
# ), f"Stream #2 index={trackDescriptors[2].getIndex()} [{trackDescriptors[2].getType().label()}:{trackDescriptors[2].getSubIndex()}] has not preserved default disposition"
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert not (trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has set default disposition"
|
||||||
|
|
||||||
|
# source subIndex 1
|
||||||
|
assert not (trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has set default disposition"
|
||||||
|
|
||||||
|
# source subIndex 2
|
||||||
|
assert not (trackDescriptors[2].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #2 index={trackDescriptors[2].getIndex()} [{trackDescriptors[2].getType().label()}:{trackDescriptors[2].getSubIndex()}] has set default disposition"
|
||||||
|
return f
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
97
src/ffx/test/disposition_combinator_3_1.py
Normal file
97
src/ffx/test/disposition_combinator_3_1.py
Normal file
@@ -0,0 +1,97 @@
|
|||||||
|
import os, sys, importlib, glob, inspect
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from .disposition_combinator_3 import DispositionCombinator3
|
||||||
|
|
||||||
|
|
||||||
|
class DispositionCombinator31(DispositionCombinator3):
|
||||||
|
|
||||||
|
VARIANT = 'D100'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return DispositionCombinator31.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
subtrack0 = set()
|
||||||
|
subtrack1 = set()
|
||||||
|
subtrack2 = set([TrackDisposition.DEFAULT])
|
||||||
|
else:
|
||||||
|
subtrack0 = set([TrackDisposition.DEFAULT])
|
||||||
|
subtrack1 = set()
|
||||||
|
subtrack2 = set()
|
||||||
|
|
||||||
|
#NOTE: Current ffmpeg version will not set most of the dispositions on arbitrary tracks
|
||||||
|
# so some checks for preserved dispositions are omitted for now
|
||||||
|
if self.__createPresets:
|
||||||
|
# subtrack0.add(TrackDisposition.COMMENT) # COMMENT
|
||||||
|
subtrack1.add(TrackDisposition.FORCED) # DESCRIPTIONS
|
||||||
|
# subtrack2.add(TrackDisposition.HEARING_IMPAIRED) # HEARING_IMPAIRED
|
||||||
|
|
||||||
|
return (subtrack0,
|
||||||
|
subtrack1,
|
||||||
|
subtrack2)
|
||||||
|
|
||||||
|
def createAssertFunc(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has not set default disposition"
|
||||||
|
# assert (trackDescriptors[0].getDispositionFlag(TrackDisposition.COMMENT)
|
||||||
|
# ), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has not preserved default disposition"
|
||||||
|
|
||||||
|
# source subIndex 1
|
||||||
|
assert (not trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has set default disposition"
|
||||||
|
assert (trackDescriptors[1].getDispositionFlag(TrackDisposition.FORCED)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has not preserved default disposition"
|
||||||
|
|
||||||
|
# source subIndex 2
|
||||||
|
assert (not trackDescriptors[2].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #2 index={trackDescriptors[2].getIndex()} [{trackDescriptors[2].getType().label()}:{trackDescriptors[2].getSubIndex()}] has set default disposition"
|
||||||
|
# assert (trackDescriptors[2].getDispositionFlag(TrackDisposition.HEARING_IMPAIRED)
|
||||||
|
# ), f"Stream #2 index={trackDescriptors[2].getIndex()} [{trackDescriptors[2].getType().label()}:{trackDescriptors[2].getSubIndex()}] has not preserved default disposition"
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has not set default disposition"
|
||||||
|
|
||||||
|
# source subIndex 1
|
||||||
|
assert (not trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has set default disposition"
|
||||||
|
|
||||||
|
# source subIndex 2
|
||||||
|
assert (not trackDescriptors[2].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #2 index={trackDescriptors[2].getIndex()} [{trackDescriptors[2].getType().label()}:{trackDescriptors[2].getSubIndex()}] has set default disposition"
|
||||||
|
return f
|
||||||
|
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
89
src/ffx/test/disposition_combinator_3_2.py
Normal file
89
src/ffx/test/disposition_combinator_3_2.py
Normal file
@@ -0,0 +1,89 @@
|
|||||||
|
import os, sys, importlib, glob, inspect
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from .disposition_combinator_3 import DispositionCombinator3
|
||||||
|
|
||||||
|
|
||||||
|
class DispositionCombinator32(DispositionCombinator3):
|
||||||
|
|
||||||
|
VARIANT = 'D010'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return DispositionCombinator32.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
subtrack0 = set([TrackDisposition.DEFAULT])
|
||||||
|
subtrack1 = set()
|
||||||
|
subtrack2 = set()
|
||||||
|
else:
|
||||||
|
subtrack0 = set()
|
||||||
|
subtrack1 = set([TrackDisposition.DEFAULT])
|
||||||
|
subtrack2 = set()
|
||||||
|
|
||||||
|
#NOTE: Current ffmpeg version will not set most of the dispositions on arbitrary tracks
|
||||||
|
# so some checks for preserved dispositions are omitted for now
|
||||||
|
if self.__createPresets:
|
||||||
|
# subtrack0.add(TrackDisposition.COMMENT) # COMMENT
|
||||||
|
# subtrack1.add(TrackDisposition.DESCRIPTIONS) # DESCRIPTIONS
|
||||||
|
subtrack2.add(TrackDisposition.FORCED) # HEARING_IMPAIRED
|
||||||
|
|
||||||
|
return (subtrack0,
|
||||||
|
subtrack1,
|
||||||
|
subtrack2)
|
||||||
|
|
||||||
|
|
||||||
|
def createAssertFunc(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (not trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has set default disposition"
|
||||||
|
# assert (trackDescriptors[0].getDispositionFlag(TrackDisposition.COMMENT)
|
||||||
|
# ), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has not preserved set default disposition"
|
||||||
|
# source subIndex 1
|
||||||
|
assert (trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has not set default disposition"
|
||||||
|
# assert (trackDescriptors[1].getDispositionFlag(TrackDisposition.DESCRIPTIONS)
|
||||||
|
# ), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has not preserved descriptions disposition"
|
||||||
|
# source subIndex 2
|
||||||
|
assert (not trackDescriptors[2].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #2 index={trackDescriptors[2].getIndex()} [{trackDescriptors[2].getType().label()}:{trackDescriptors[2].getSubIndex()}] has set default disposition"
|
||||||
|
assert (trackDescriptors[2].getDispositionFlag(TrackDisposition.FORCED)
|
||||||
|
), f"Stream #2 index={trackDescriptors[2].getIndex()} [{trackDescriptors[2].getType().label()}:{trackDescriptors[2].getSubIndex()}] has not preserved default disposition"
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (not trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has set default disposition"
|
||||||
|
# source subIndex 1
|
||||||
|
assert (trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has not set default disposition"
|
||||||
|
# source subIndex 2
|
||||||
|
assert (not trackDescriptors[2].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #2 index={trackDescriptors[2].getIndex()} [{trackDescriptors[2].getType().label()}:{trackDescriptors[2].getSubIndex()}] has set default disposition"
|
||||||
|
return f
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
89
src/ffx/test/disposition_combinator_3_3.py
Normal file
89
src/ffx/test/disposition_combinator_3_3.py
Normal file
@@ -0,0 +1,89 @@
|
|||||||
|
import os, sys, importlib, glob, inspect
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from .disposition_combinator_3 import DispositionCombinator3
|
||||||
|
|
||||||
|
|
||||||
|
class DispositionCombinator33(DispositionCombinator3):
|
||||||
|
|
||||||
|
VARIANT = 'D001'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return DispositionCombinator33.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
subtrack0 = set()
|
||||||
|
subtrack1 = set([TrackDisposition.DEFAULT])
|
||||||
|
subtrack2 = set()
|
||||||
|
else:
|
||||||
|
subtrack0 = set()
|
||||||
|
subtrack1 = set()
|
||||||
|
subtrack2 = set([TrackDisposition.DEFAULT])
|
||||||
|
|
||||||
|
#NOTE: Current ffmpeg version will not set most of the dispositions on arbitrary tracks
|
||||||
|
# so some checks for preserved dispositions are omitted for now
|
||||||
|
if self.__createPresets:
|
||||||
|
# subtrack0.add(TrackDisposition.COMMENT) # COMMENT
|
||||||
|
subtrack1.add(TrackDisposition.FORCED) # DESCRIPTIONS
|
||||||
|
# subtrack2.add(TrackDisposition.HEARING_IMPAIRED) # HEARING_IMPAIRED
|
||||||
|
|
||||||
|
return (subtrack0,
|
||||||
|
subtrack1,
|
||||||
|
subtrack2)
|
||||||
|
|
||||||
|
|
||||||
|
def createAssertFunc(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (not trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has set default disposition"
|
||||||
|
# assert (trackDescriptors[0].getDispositionFlag(TrackDisposition.COMMENT)
|
||||||
|
# ), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has not preserved set default disposition"
|
||||||
|
# source subIndex 1
|
||||||
|
assert (not trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has set default disposition"
|
||||||
|
assert (trackDescriptors[1].getDispositionFlag(TrackDisposition.FORCED)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has not preserved descriptions disposition"
|
||||||
|
# source subIndex 2
|
||||||
|
assert (trackDescriptors[2].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #2 index={trackDescriptors[2].getIndex()} [{trackDescriptors[2].getType().label()}:{trackDescriptors[2].getSubIndex()}] has not set default disposition"
|
||||||
|
# assert (trackDescriptors[2].getDispositionFlag(TrackDisposition.HEARING_IMPAIRED)
|
||||||
|
# ), f"Stream #2 index={trackDescriptors[2].getIndex()} [{trackDescriptors[2].getType().label()}:{trackDescriptors[2].getSubIndex()}] has not preserved default disposition"
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
def f(assertObj: dict):
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
trackDescriptors = assertObj['tracks']
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (not trackDescriptors[0].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #0 index={trackDescriptors[0].getIndex()} [{trackDescriptors[0].getType().label()}:{trackDescriptors[0].getSubIndex()}] has set default disposition"
|
||||||
|
# source subIndex 1
|
||||||
|
assert (not trackDescriptors[1].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #1 index={trackDescriptors[1].getIndex()} [{trackDescriptors[1].getType().label()}:{trackDescriptors[1].getSubIndex()}] has set default disposition"
|
||||||
|
# source subIndex 2
|
||||||
|
assert (trackDescriptors[2].getDispositionFlag(TrackDisposition.DEFAULT)
|
||||||
|
), f"Stream #2 index={trackDescriptors[2].getIndex()} [{trackDescriptors[2].getType().label()}:{trackDescriptors[2].getSubIndex()}] has not set default disposition"
|
||||||
|
return f
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
46
src/ffx/test/disposition_combinator_3_4.py
Normal file
46
src/ffx/test/disposition_combinator_3_4.py
Normal file
@@ -0,0 +1,46 @@
|
|||||||
|
import os, sys, importlib, glob, inspect
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from .disposition_combinator_3 import DispositionCombinator3
|
||||||
|
|
||||||
|
|
||||||
|
class DispositionCombinator34(DispositionCombinator3):
|
||||||
|
|
||||||
|
VARIANT = 'D101'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return DispositionCombinator34.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
|
||||||
|
subtrack0 = set([TrackDisposition.DEFAULT])
|
||||||
|
subtrack1 = set()
|
||||||
|
subtrack2 = set([TrackDisposition.DEFAULT])
|
||||||
|
|
||||||
|
#NOTE: Current ffmpeg version will not set most of the dispositions on arbitrary tracks
|
||||||
|
# so some checks for preserved dispositions are omitted for now
|
||||||
|
if self.__createPresets:
|
||||||
|
subtrack0.add(TrackDisposition.FORCED) # COMMENT
|
||||||
|
# subtrack1.add(TrackDisposition.DESCRIPTIONS) # DESCRIPTIONS
|
||||||
|
# subtrack2.add(TrackDisposition.HEARING_IMPAIRED) # HEARING_IMPAIRED
|
||||||
|
|
||||||
|
return (subtrack0,
|
||||||
|
subtrack1,
|
||||||
|
subtrack2)
|
||||||
|
|
||||||
|
#TODO: tmpdb cases
|
||||||
|
def createAssertFunc(self):
|
||||||
|
def f(assertObj: dict = {}):
|
||||||
|
pass
|
||||||
|
return f
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return True
|
||||||
278
src/ffx/test/helper.py
Normal file
278
src/ffx/test/helper.py
Normal file
@@ -0,0 +1,278 @@
|
|||||||
|
import os, math, tempfile, click
|
||||||
|
|
||||||
|
|
||||||
|
from ffx.ffx_controller import FfxController
|
||||||
|
|
||||||
|
from ffx.process import executeProcess
|
||||||
|
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.helper import dictCache
|
||||||
|
from ffx.configuration_controller import ConfigurationController
|
||||||
|
|
||||||
|
|
||||||
|
SHORT_SUBTITLE_SEQUENCE = [{'start': 1, 'end': 2, 'text': 'yolo'},
|
||||||
|
{'start': 3, 'end': 4, 'text': 'zolo'},
|
||||||
|
{'start': 5, 'end': 6, 'text': 'golo'}]
|
||||||
|
|
||||||
|
def getTimeString(hours: float = 0.0,
|
||||||
|
minutes: float = 0.0,
|
||||||
|
seconds: float = 0.0,
|
||||||
|
millis: float = 0.0,
|
||||||
|
format: str = ''):
|
||||||
|
|
||||||
|
duration = (hours * 3600.0
|
||||||
|
+ minutes * 60.0
|
||||||
|
+ seconds
|
||||||
|
+ millis / 1000.0)
|
||||||
|
|
||||||
|
hours = math.floor(duration / 3600.0)
|
||||||
|
remaining = duration - 3600.0 * hours
|
||||||
|
|
||||||
|
minutes = math.floor(remaining / 60.0)
|
||||||
|
remaining = remaining - 60.0 * minutes
|
||||||
|
|
||||||
|
seconds = math.floor(remaining)
|
||||||
|
remaining = remaining - seconds
|
||||||
|
|
||||||
|
millis = math.floor(remaining * 1000)
|
||||||
|
|
||||||
|
if format == 'ass':
|
||||||
|
return f"{hours:01d}:{minutes:02d}:{seconds:02d}.{millis:02d}"
|
||||||
|
|
||||||
|
# srt & vtt
|
||||||
|
return f"{hours:02d}:{minutes:02d}:{seconds:02d}.{millis:03d}"
|
||||||
|
|
||||||
|
|
||||||
|
def createAssFile(entries: dict, directory = None):
|
||||||
|
|
||||||
|
# [Script Info]
|
||||||
|
# ; Script generated by FFmpeg/Lavc61.3.100
|
||||||
|
# ScriptType: v4.00+
|
||||||
|
# PlayResX: 384
|
||||||
|
# PlayResY: 288
|
||||||
|
# ScaledBorderAndShadow: yes
|
||||||
|
# YCbCr Matrix: None
|
||||||
|
#
|
||||||
|
# [V4+ Styles]
|
||||||
|
# Format: Name, Fontname, Fontsize, PrimaryColour, SecondaryColour, OutlineColour, BackColour, Bold, Italic, Underline, StrikeOut, ScaleX, ScaleY, Spacing, Angle, BorderStyle, Outline, Shadow, Alignment, MarginL, MarginR, MarginV, Encoding
|
||||||
|
# Style: Default,Arial,16,&Hffffff,&Hffffff,&H0,&H0,0,0,0,0,100,100,0,0,1,1,0,2,10,10,10,1
|
||||||
|
#
|
||||||
|
# [Events]
|
||||||
|
# Format: Layer, Start, End, Style, Name, MarginL, MarginR, MarginV, Effect, Text
|
||||||
|
# Dialogue: 0,0:00:01.00,0:00:02.00,Default,,0,0,0,,yolo
|
||||||
|
# Dialogue: 0,0:00:03.00,0:00:04.00,Default,,0,0,0,,zolo
|
||||||
|
# Dialogue: 0,0:00:05.00,0:00:06.00,Default,,0,0,0,,golo
|
||||||
|
tmpFileName = tempfile.mktemp(suffix=".ass", dir = directory)
|
||||||
|
|
||||||
|
with open(tmpFileName, 'w') as tmpFile:
|
||||||
|
|
||||||
|
tmpFile.write("[Script Info]\n")
|
||||||
|
tmpFile.write("; Script generated by Ffx\n")
|
||||||
|
tmpFile.write("ScriptType: v4.00+\n")
|
||||||
|
tmpFile.write("PlayResX: 384\n")
|
||||||
|
tmpFile.write("PlayResY: 288\n")
|
||||||
|
tmpFile.write("ScaledBorderAndShadow: yes\n")
|
||||||
|
tmpFile.write("YCbCr Matrix: None\n")
|
||||||
|
tmpFile.write("\n")
|
||||||
|
tmpFile.write("[V4+ Styles]\n")
|
||||||
|
tmpFile.write("Format: Name, Fontname, Fontsize, PrimaryColour, SecondaryColour, OutlineColour, BackColour, Bold, Italic, Underline, StrikeOut, ScaleX, ScaleY, Spacing, Angle, BorderStyle, Outline, Shadow, Alignment, MarginL, MarginR, MarginV, Encoding\n")
|
||||||
|
tmpFile.write("Style: Default,Arial,16,&Hffffff,&Hffffff,&H0,&H0,0,0,0,0,100,100,0,0,1,1,0,2,10,10,10,1\n")
|
||||||
|
tmpFile.write("\n")
|
||||||
|
tmpFile.write("[Events]\n")
|
||||||
|
tmpFile.write("Format: Layer, Start, End, Style, Name, MarginL, MarginR, MarginV, Effect, Text\n")
|
||||||
|
|
||||||
|
for entryIndex in range(len(entries)):
|
||||||
|
tmpFile.write(f"Dialogue: 0,{getTimeString(seconds=entries[entryIndex]['start'], format='ass')},{getTimeString(seconds=entries[entryIndex]['end'], format='ass')},Default,,0,0,0,,{entries[entryIndex]['text']}\n")
|
||||||
|
|
||||||
|
return tmpFileName
|
||||||
|
|
||||||
|
def createSrtFile(entries: dict, directory = None):
|
||||||
|
# 1
|
||||||
|
# 00:00:00,000 --> 00:00:02,500
|
||||||
|
# Welcome to the Example Subtitle File!
|
||||||
|
#
|
||||||
|
# 2
|
||||||
|
# 00:00:03,000 --> 00:00:06,000
|
||||||
|
# This is a demonstration of SRT subtitles.
|
||||||
|
#
|
||||||
|
# 3
|
||||||
|
# 00:00:07,000 --> 00:00:10,500
|
||||||
|
# You can use SRT files to add subtitles to your videos.
|
||||||
|
|
||||||
|
tmpFileName = tempfile.mktemp(suffix=".srt", dir = directory)
|
||||||
|
|
||||||
|
with open(tmpFileName, 'w') as tmpFile:
|
||||||
|
|
||||||
|
for entryIndex in range(len(entries)):
|
||||||
|
|
||||||
|
tmpFile.write(f"{entryIndex}\n")
|
||||||
|
tmpFile.write(f"{getTimeString(seconds=entries[entryIndex]['start'])} --> {getTimeString(seconds=entries[entryIndex]['end'])}\n")
|
||||||
|
tmpFile.write(f"{entries[entryIndex]['text']}\n\n")
|
||||||
|
|
||||||
|
return tmpFileName
|
||||||
|
|
||||||
|
|
||||||
|
def createVttFile(entries: dict, directory = None):
|
||||||
|
# WEBVTT
|
||||||
|
#
|
||||||
|
# 01:20:33.050 --> 01:20:35.050
|
||||||
|
# Yolo
|
||||||
|
|
||||||
|
tmpFileName = tempfile.mktemp(suffix=".vtt", dir = directory)
|
||||||
|
|
||||||
|
with open(tmpFileName, 'w') as tmpFile:
|
||||||
|
|
||||||
|
tmpFile.write("WEBVTT\n")
|
||||||
|
|
||||||
|
for entryIndex in range(len(entries)):
|
||||||
|
|
||||||
|
tmpFile.write("\n")
|
||||||
|
tmpFile.write(f"{getTimeString(seconds=entries[entryIndex]['start'])} --> {getTimeString(seconds=entries[entryIndex]['end'])}\n")
|
||||||
|
tmpFile.write(f"{entries[entryIndex]['text']}\n")
|
||||||
|
|
||||||
|
|
||||||
|
return tmpFileName
|
||||||
|
|
||||||
|
|
||||||
|
def createMediaTestFile(mediaDescriptor: MediaDescriptor,
|
||||||
|
directory: str = '',
|
||||||
|
baseName: str = 'media',
|
||||||
|
format: str = '',
|
||||||
|
extension: str = 'mkv',
|
||||||
|
sizeX: int = 1280,
|
||||||
|
sizeY: int = 720,
|
||||||
|
rate: int = 25,
|
||||||
|
length: int = 10,
|
||||||
|
logger = None):
|
||||||
|
|
||||||
|
# subtitleFilePath = createVttFile(SHORT_SUBTITLE_SEQUENCE)
|
||||||
|
|
||||||
|
# commandTokens = FfxController.COMMAND_TOKENS
|
||||||
|
commandTokens = ['ffmpeg', '-y']
|
||||||
|
|
||||||
|
generatorCache = []
|
||||||
|
generatorTokens = []
|
||||||
|
mappingTokens = []
|
||||||
|
importTokens = []
|
||||||
|
metadataTokens = []
|
||||||
|
|
||||||
|
|
||||||
|
for mediaTagKey, mediaTagValue in mediaDescriptor.getTags().items():
|
||||||
|
metadataTokens += ['-metadata:g', f"{mediaTagKey}={mediaTagValue}"]
|
||||||
|
|
||||||
|
subIndexCounter = {}
|
||||||
|
|
||||||
|
for trackDescriptor in mediaDescriptor.getAllTrackDescriptors():
|
||||||
|
|
||||||
|
trackType = trackDescriptor.getType()
|
||||||
|
|
||||||
|
if trackType == TrackType.VIDEO:
|
||||||
|
|
||||||
|
cacheIndex, generatorCache = dictCache({'type': TrackType.VIDEO}, generatorCache)
|
||||||
|
# click.echo(f"createMediaTestFile() cache index={cacheIndex} size={len(generatorCache)}")
|
||||||
|
|
||||||
|
if cacheIndex == -1:
|
||||||
|
generatorTokens += ['-f',
|
||||||
|
'lavfi',
|
||||||
|
'-i',
|
||||||
|
f"color=size={sizeX}x{sizeY}:rate={rate}:color=black"]
|
||||||
|
|
||||||
|
sourceIndex = len(generatorCache) - 1 if cacheIndex == -1 else cacheIndex
|
||||||
|
mappingTokens += ['-map', f"{sourceIndex}:v:0"]
|
||||||
|
|
||||||
|
if not trackType in subIndexCounter.keys():
|
||||||
|
subIndexCounter[trackType] = 0
|
||||||
|
for mediaTagKey, mediaTagValue in trackDescriptor.getTags().items():
|
||||||
|
metadataTokens += [f"-metadata:s:{trackType.indicator()}:{subIndexCounter[trackType]}",
|
||||||
|
f"{mediaTagKey}={mediaTagValue}"]
|
||||||
|
subIndexCounter[trackType] += 1
|
||||||
|
|
||||||
|
if trackType == TrackType.AUDIO:
|
||||||
|
|
||||||
|
audioLayout = 'stereo'
|
||||||
|
|
||||||
|
cacheIndex, generatorCache = dictCache({'type': TrackType.AUDIO, 'layout': audioLayout}, generatorCache)
|
||||||
|
# click.echo(f"createMediaTestFile() cache index={cacheIndex} size={len(generatorCache)}")
|
||||||
|
|
||||||
|
# click.echo(f"generartorCache index={cacheIndex} len={len(generatorCache)}")
|
||||||
|
if cacheIndex == -1:
|
||||||
|
generatorTokens += ['-f',
|
||||||
|
'lavfi',
|
||||||
|
'-i',
|
||||||
|
f"anullsrc=channel_layout={audioLayout}:sample_rate=44100"]
|
||||||
|
|
||||||
|
sourceIndex = len(generatorCache) - 1 if cacheIndex == -1 else cacheIndex
|
||||||
|
mappingTokens += ['-map', f"{sourceIndex}:a:0"]
|
||||||
|
|
||||||
|
if not trackType in subIndexCounter.keys():
|
||||||
|
subIndexCounter[trackType] = 0
|
||||||
|
for mediaTagKey, mediaTagValue in trackDescriptor.getTags().items():
|
||||||
|
metadataTokens += [f"-metadata:s:{trackType.indicator()}:{subIndexCounter[trackType]}",
|
||||||
|
f"{mediaTagKey}={mediaTagValue}"]
|
||||||
|
subIndexCounter[trackType] += 1
|
||||||
|
|
||||||
|
if trackType == TrackType.SUBTITLE:
|
||||||
|
|
||||||
|
cacheIndex, generatorCache = dictCache({'type': TrackType.SUBTITLE}, generatorCache)
|
||||||
|
# click.echo(f"createMediaTestFile() cache index={cacheIndex} size={len(generatorCache)}")
|
||||||
|
|
||||||
|
if cacheIndex == -1:
|
||||||
|
importTokens = ['-i', createVttFile(SHORT_SUBTITLE_SEQUENCE, directory=directory if directory else None)]
|
||||||
|
|
||||||
|
sourceIndex = len(generatorCache) - 1 if cacheIndex == -1 else cacheIndex
|
||||||
|
mappingTokens += ['-map', f"{sourceIndex}:s:0"]
|
||||||
|
|
||||||
|
if not trackType in subIndexCounter.keys():
|
||||||
|
subIndexCounter[trackType] = 0
|
||||||
|
for mediaTagKey, mediaTagValue in trackDescriptor.getTags().items():
|
||||||
|
metadataTokens += [f"-metadata:s:{trackType.indicator()}:{subIndexCounter[trackType]}",
|
||||||
|
f"{mediaTagKey}={mediaTagValue}"]
|
||||||
|
subIndexCounter[trackType] += 1
|
||||||
|
|
||||||
|
#TODO: Optimize too many runs
|
||||||
|
ffxContext = {'config': ConfigurationController(), 'logger': logger}
|
||||||
|
fc = FfxController(ffxContext, mediaDescriptor)
|
||||||
|
|
||||||
|
commandTokens += (generatorTokens
|
||||||
|
+ importTokens
|
||||||
|
+ mappingTokens
|
||||||
|
+ metadataTokens
|
||||||
|
+ fc.generateDispositionTokens())
|
||||||
|
|
||||||
|
|
||||||
|
commandTokens += ['-t', str(length)]
|
||||||
|
|
||||||
|
if format:
|
||||||
|
commandTokens += ['-f', format]
|
||||||
|
|
||||||
|
fileName = f"{baseName}.{extension}"
|
||||||
|
|
||||||
|
if directory:
|
||||||
|
outputPath = os.path.join(directory, fileName)
|
||||||
|
else:
|
||||||
|
outputPath = fileName
|
||||||
|
|
||||||
|
commandTokens += [outputPath]
|
||||||
|
|
||||||
|
|
||||||
|
ctx = {'logger': logger}
|
||||||
|
|
||||||
|
out, err, rc = executeProcess(commandTokens, context = ctx)
|
||||||
|
|
||||||
|
if not logger is None:
|
||||||
|
if out:
|
||||||
|
logger.debug(f"createMediaTestFile(): Process output: {out}")
|
||||||
|
if rc:
|
||||||
|
logger.debug(f"createMediaTestFile(): Process returned ERROR {rc} ({err})")
|
||||||
|
|
||||||
|
|
||||||
|
return outputPath
|
||||||
|
|
||||||
|
|
||||||
|
def createEmptyDirectory():
|
||||||
|
return tempfile.mkdtemp()
|
||||||
|
|
||||||
|
def createEmptyFile(suffix=None):
|
||||||
|
return tempfile.mkstemp(suffix=suffix)
|
||||||
43
src/ffx/test/indicator_combinator.py
Normal file
43
src/ffx/test/indicator_combinator.py
Normal file
@@ -0,0 +1,43 @@
|
|||||||
|
class IndicatorCombinator():
|
||||||
|
|
||||||
|
IDENTIFIER = 'indicator'
|
||||||
|
|
||||||
|
MAX_SEASON = 2
|
||||||
|
MAX_EPISODE = 3
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getIdentifier(self):
|
||||||
|
return IndicatorCombinator.IDENTIFIER
|
||||||
|
|
||||||
|
def getPayload(self, season: int = -1, episode: int = -1):
|
||||||
|
if season == -1 and episode == -1:
|
||||||
|
return {
|
||||||
|
'variant': 'S00E00',
|
||||||
|
'indicator': '',
|
||||||
|
'season': season,
|
||||||
|
'episode': episode
|
||||||
|
}
|
||||||
|
else:
|
||||||
|
return {
|
||||||
|
'variant': f"S{season:02d}E{episode:02d}",
|
||||||
|
'indicator': f"S{season:02d}E{episode:02d}",
|
||||||
|
'season': season,
|
||||||
|
'episode': episode
|
||||||
|
}
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
|
||||||
|
yield self.getPayload()
|
||||||
|
for season in range(IndicatorCombinator.MAX_SEASON):
|
||||||
|
for episode in range(IndicatorCombinator.MAX_EPISODE):
|
||||||
|
yield self.getPayload(season+1, episode+1)
|
||||||
36
src/ffx/test/label_combinator.py
Normal file
36
src/ffx/test/label_combinator.py
Normal file
@@ -0,0 +1,36 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
class LabelCombinator():
|
||||||
|
|
||||||
|
IDENTIFIER = 'label'
|
||||||
|
PREFIX = 'label_combinator_'
|
||||||
|
|
||||||
|
LABEL = 'ffx'
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getIdentifier(self):
|
||||||
|
return LabelCombinator.IDENTIFIER
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def list():
|
||||||
|
basePath = os.path.dirname(__file__)
|
||||||
|
return [os.path.basename(p)[len(LabelCombinator.PREFIX):-3]
|
||||||
|
for p
|
||||||
|
in glob.glob(f"{ basePath }/{LabelCombinator.PREFIX}*.py", recursive = True)
|
||||||
|
if p != __file__]
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getClassReference(identifier):
|
||||||
|
importlib.import_module(f"ffx.test.{LabelCombinator.PREFIX}{ identifier }")
|
||||||
|
for name, obj in inspect.getmembers(sys.modules[f"ffx.test.{LabelCombinator.PREFIX}{ identifier }"]):
|
||||||
|
#HINT: Excluding MediaCombinator as it seems to be included by import (?)
|
||||||
|
if inspect.isclass(obj) and name != 'LabelCombinator' and name.startswith('LabelCombinator'):
|
||||||
|
return obj
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getAllClassReferences():
|
||||||
|
return [LabelCombinator.getClassReference(i) for i in LabelCombinator.list()]
|
||||||
30
src/ffx/test/label_combinator_0.py
Normal file
30
src/ffx/test/label_combinator_0.py
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .label_combinator import LabelCombinator
|
||||||
|
|
||||||
|
|
||||||
|
class LabelCombinator0(LabelCombinator):
|
||||||
|
|
||||||
|
VARIANT = 'L0'
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return LabelCombinator0.VARIANT
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
return ''
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
30
src/ffx/test/label_combinator_1.py
Normal file
30
src/ffx/test/label_combinator_1.py
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .label_combinator import LabelCombinator
|
||||||
|
|
||||||
|
class LabelCombinator1(LabelCombinator):
|
||||||
|
|
||||||
|
VARIANT = 'L1'
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return LabelCombinator1.VARIANT
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
return LabelCombinator.LABEL
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
33
src/ffx/test/media_combinator.py
Normal file
33
src/ffx/test/media_combinator.py
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
class MediaCombinator():
|
||||||
|
|
||||||
|
IDENTIFIER = 'media'
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getIdentifier(self):
|
||||||
|
return MediaCombinator.IDENTIFIER
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def list():
|
||||||
|
basePath = os.path.dirname(__file__)
|
||||||
|
return [os.path.basename(p)[17:-3]
|
||||||
|
for p
|
||||||
|
in glob.glob(f"{ basePath }/media_combinator_*.py", recursive = True)
|
||||||
|
if p != __file__]
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getClassReference(identifier):
|
||||||
|
importlib.import_module(f"ffx.test.media_combinator_{ identifier }")
|
||||||
|
for name, obj in inspect.getmembers(sys.modules[f"ffx.test.media_combinator_{ identifier }"]):
|
||||||
|
#HINT: Excluding MediaCombinator as it seems to be included by import (?)
|
||||||
|
if inspect.isclass(obj) and name != 'MediaCombinator' and name.startswith('MediaCombinator'):
|
||||||
|
return obj
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getAllClassReferences():
|
||||||
|
return [MediaCombinator.getClassReference(i) for i in MediaCombinator.list()]
|
||||||
103
src/ffx/test/media_combinator_0.py
Normal file
103
src/ffx/test/media_combinator_0.py
Normal file
@@ -0,0 +1,103 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .media_combinator import MediaCombinator
|
||||||
|
from .media_tag_combinator import MediaTagCombinator
|
||||||
|
|
||||||
|
|
||||||
|
class MediaCombinator0(MediaCombinator):
|
||||||
|
|
||||||
|
VARIANT = 'VA'
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return MediaCombinator0.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.VIDEO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor1 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.AUDIO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor2 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[MediaDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY] = [trackDescriptor1,
|
||||||
|
trackDescriptor2]
|
||||||
|
mediaDescriptor = MediaDescriptor(**kwargs)
|
||||||
|
# mediaDescriptor.reindexSubIndices()
|
||||||
|
|
||||||
|
return mediaDescriptor
|
||||||
|
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
|
||||||
|
for MTC in MediaTagCombinator.getAllClassReferences():
|
||||||
|
|
||||||
|
mtc = MTC(self._context)
|
||||||
|
|
||||||
|
yObj = {}
|
||||||
|
|
||||||
|
yObj['identifier'] = self.getIdentifier()
|
||||||
|
yObj['variants'] = [self.getVariant(),
|
||||||
|
mtc.getVariant()]
|
||||||
|
yObj['payload'] = self.getPayload()
|
||||||
|
|
||||||
|
yObj['assertSelectors'] = ['M', 'MT']
|
||||||
|
yObj['assertFuncs'] = [self.assertFunc,
|
||||||
|
mtc.createAssertFunc()]
|
||||||
|
yObj['shouldFail'] = (self.shouldFail()
|
||||||
|
| mtc.shouldFail())
|
||||||
|
|
||||||
|
yieldObj = {'target': yObj}
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
mtc_p = MTC(self._context, createPresets = True)
|
||||||
|
|
||||||
|
yObj_p = {}
|
||||||
|
|
||||||
|
yObj_p['identifier'] = self.getIdentifier()
|
||||||
|
yObj_p['variants'] = [self.getVariant(),
|
||||||
|
mtc_p.getVariant()]
|
||||||
|
|
||||||
|
yObj_p['payload'] = self.getPayload()
|
||||||
|
|
||||||
|
yObj_p['assertSelectors'] = ['M', 'MT']
|
||||||
|
yObj_p['assertFuncs'] = [self.assertFunc,
|
||||||
|
mtc_p.createAssertFunc()]
|
||||||
|
yObj_p['shouldFail'] = (self.shouldFail()
|
||||||
|
| mtc_p.shouldFail())
|
||||||
|
|
||||||
|
yieldObj['preset'] = yObj_p
|
||||||
|
|
||||||
|
yield yieldObj
|
||||||
114
src/ffx/test/media_combinator_1.py
Normal file
114
src/ffx/test/media_combinator_1.py
Normal file
@@ -0,0 +1,114 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .media_combinator import MediaCombinator
|
||||||
|
from .media_tag_combinator import MediaTagCombinator
|
||||||
|
|
||||||
|
|
||||||
|
class MediaCombinator1(MediaCombinator):
|
||||||
|
|
||||||
|
VARIANT = 'VAS'
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return MediaCombinator1.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.VIDEO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor0 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.AUDIO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor1 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.SUBTITLE
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor2 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[MediaDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY] = [trackDescriptor0,
|
||||||
|
trackDescriptor1,
|
||||||
|
trackDescriptor2]
|
||||||
|
|
||||||
|
mediaDescriptor = MediaDescriptor(**kwargs)
|
||||||
|
# mediaDescriptor.reindexSubIndices()
|
||||||
|
|
||||||
|
return mediaDescriptor
|
||||||
|
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
|
||||||
|
for MTC in MediaTagCombinator.getAllClassReferences():
|
||||||
|
|
||||||
|
mtc = MTC(self._context)
|
||||||
|
|
||||||
|
yObj = {}
|
||||||
|
|
||||||
|
yObj['identifier'] = self.getIdentifier()
|
||||||
|
yObj['variants'] = [self.getVariant(),
|
||||||
|
mtc.getVariant()]
|
||||||
|
|
||||||
|
yObj['payload'] = self.getPayload()
|
||||||
|
|
||||||
|
yObj['assertSelectors'] = ['M', 'MT']
|
||||||
|
yObj['assertFuncs'] = [self.assertFunc,
|
||||||
|
mtc.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj['shouldFail'] = (self.shouldFail()
|
||||||
|
| mtc.shouldFail())
|
||||||
|
|
||||||
|
yieldObj = {'target': yObj}
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
mtc_p = MTC(self._context, createPresets = True)
|
||||||
|
|
||||||
|
yObj_p = {}
|
||||||
|
|
||||||
|
yObj_p['identifier'] = self.getIdentifier()
|
||||||
|
yObj_p['variants'] = [self.getVariant(),
|
||||||
|
mtc_p.getVariant()]
|
||||||
|
|
||||||
|
yObj_p['payload'] = self.getPayload()
|
||||||
|
|
||||||
|
yObj_p['assertSelectors'] = ['M', 'MT']
|
||||||
|
yObj_p['assertFuncs'] = [self.assertFunc,
|
||||||
|
mtc_p.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj_p['shouldFail'] = (self.shouldFail()
|
||||||
|
| mtc_p.shouldFail())
|
||||||
|
yieldObj['preset'] = yObj_p
|
||||||
|
|
||||||
|
yield yieldObj
|
||||||
156
src/ffx/test/media_combinator_2.py
Normal file
156
src/ffx/test/media_combinator_2.py
Normal file
@@ -0,0 +1,156 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools, click
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .media_combinator import MediaCombinator
|
||||||
|
|
||||||
|
from .disposition_combinator_2 import DispositionCombinator2
|
||||||
|
from .track_tag_combinator_2 import TrackTagCombinator2
|
||||||
|
from .permutation_combinator_2 import PermutationCombinator2
|
||||||
|
from .media_tag_combinator import MediaTagCombinator
|
||||||
|
|
||||||
|
|
||||||
|
class MediaCombinator2(MediaCombinator):
|
||||||
|
|
||||||
|
VARIANT = 'VASS'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return MediaCombinator2.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self,
|
||||||
|
subtitleDispositionTuple = (set(), set()),
|
||||||
|
subtitleTagTuple = ({}, {})):
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.VIDEO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor0 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.AUDIO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor1 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.SUBTITLE
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = subtitleDispositionTuple[0]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = subtitleTagTuple[0]
|
||||||
|
trackDescriptor2 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 3
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 3
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.SUBTITLE
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = subtitleDispositionTuple[1]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = subtitleTagTuple[1]
|
||||||
|
trackDescriptor3 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[MediaDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY] = [trackDescriptor0,
|
||||||
|
trackDescriptor1,
|
||||||
|
trackDescriptor2,
|
||||||
|
trackDescriptor3]
|
||||||
|
|
||||||
|
mediaDescriptor = MediaDescriptor(**kwargs)
|
||||||
|
# mediaDescriptor.reindexSubIndices()
|
||||||
|
|
||||||
|
return mediaDescriptor
|
||||||
|
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
|
||||||
|
for MTC in MediaTagCombinator.getAllClassReferences():
|
||||||
|
for DC2 in DispositionCombinator2.getAllClassReferences():
|
||||||
|
for TC2 in TrackTagCombinator2.getAllClassReferences():
|
||||||
|
|
||||||
|
dc2 = DC2(self._context)
|
||||||
|
tc2 = TC2(self._context)
|
||||||
|
|
||||||
|
mtc = MTC(self._context)
|
||||||
|
|
||||||
|
yObj = {}
|
||||||
|
|
||||||
|
yObj['identifier'] = self.getIdentifier()
|
||||||
|
yObj['variants'] = [self.getVariant(),
|
||||||
|
f"S:{dc2.getVariant()}",
|
||||||
|
f"S:{tc2.getVariant()}",
|
||||||
|
mtc.getVariant()]
|
||||||
|
|
||||||
|
yObj['payload'] = self.getPayload(dc2.getPayload(),
|
||||||
|
tc2.getPayload())
|
||||||
|
|
||||||
|
yObj['assertSelectors'] = ['M', 'SD', 'ST', 'MT']
|
||||||
|
yObj['assertFuncs'] = [self.assertFunc,
|
||||||
|
dc2.createAssertFunc(),
|
||||||
|
tc2.createAssertFunc(),
|
||||||
|
mtc.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj['shouldFail'] = (self.shouldFail()
|
||||||
|
| dc2.shouldFail()
|
||||||
|
| tc2.shouldFail()
|
||||||
|
| mtc.shouldFail())
|
||||||
|
|
||||||
|
yieldObj = {'target': yObj}
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
dc2_p = DC2(self._context, createPresets = True)
|
||||||
|
tc2_p = TC2(self._context, createPresets = True)
|
||||||
|
|
||||||
|
mtc_p = MTC(self._context, createPresets = True)
|
||||||
|
|
||||||
|
yObj_p = {}
|
||||||
|
|
||||||
|
yObj_p['identifier'] = self.getIdentifier()
|
||||||
|
yObj_p['variants'] = [self.getVariant(),
|
||||||
|
f"S:{dc2_p.getVariant()}",
|
||||||
|
f"S:{tc2_p.getVariant()}",
|
||||||
|
mtc_p.getVariant()]
|
||||||
|
|
||||||
|
yObj_p['payload'] = self.getPayload(dc2_p.getPayload(),
|
||||||
|
tc2_p.getPayload())
|
||||||
|
|
||||||
|
yObj_p['assertSelectors'] = ['M', 'SD', 'ST', 'MT']
|
||||||
|
yObj_p['assertFuncs'] = [self.assertFunc,
|
||||||
|
dc2_p.createAssertFunc(),
|
||||||
|
tc2_p.createAssertFunc(),
|
||||||
|
mtc_p.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj_p['shouldFail'] = (self.shouldFail()
|
||||||
|
| dc2_p.shouldFail()
|
||||||
|
| tc2_p.shouldFail()
|
||||||
|
| mtc_p.shouldFail())
|
||||||
|
|
||||||
|
yieldObj['preset'] = yObj_p
|
||||||
|
|
||||||
|
yield yieldObj
|
||||||
165
src/ffx/test/media_combinator_3.py
Normal file
165
src/ffx/test/media_combinator_3.py
Normal file
@@ -0,0 +1,165 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .media_combinator import MediaCombinator
|
||||||
|
|
||||||
|
from .disposition_combinator_3 import DispositionCombinator3
|
||||||
|
from .track_tag_combinator_3 import TrackTagCombinator3
|
||||||
|
from .permutation_combinator_3 import PermutationCombinator3
|
||||||
|
from .media_tag_combinator import MediaTagCombinator
|
||||||
|
|
||||||
|
|
||||||
|
class MediaCombinator3(MediaCombinator):
|
||||||
|
|
||||||
|
VARIANT = 'VASSS'
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return MediaCombinator3.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self,
|
||||||
|
subtitleDispositionTuple = (set(), set(), set()),
|
||||||
|
subtitleTagTuple = ({}, {}, {})):
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.VIDEO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor0 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.AUDIO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor1 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.SUBTITLE
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = subtitleDispositionTuple[0]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = subtitleTagTuple[0]
|
||||||
|
trackDescriptor2 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 3
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 3
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.SUBTITLE
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = subtitleDispositionTuple[1]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = subtitleTagTuple[1]
|
||||||
|
trackDescriptor3 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 4
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 4
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.SUBTITLE
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = subtitleDispositionTuple[2]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = subtitleTagTuple[2]
|
||||||
|
trackDescriptor4 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[MediaDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY] = [trackDescriptor0,
|
||||||
|
trackDescriptor1,
|
||||||
|
trackDescriptor2,
|
||||||
|
trackDescriptor3,
|
||||||
|
trackDescriptor4]
|
||||||
|
|
||||||
|
mediaDescriptor = MediaDescriptor(**kwargs)
|
||||||
|
# mediaDescriptor.reindexSubIndices()
|
||||||
|
|
||||||
|
return mediaDescriptor
|
||||||
|
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
|
||||||
|
for MTC in MediaTagCombinator.getAllClassReferences():
|
||||||
|
for DC3 in DispositionCombinator3.getAllClassReferences():
|
||||||
|
for TC3 in TrackTagCombinator3.getAllClassReferences():
|
||||||
|
|
||||||
|
dc3 = DC3(self._context)
|
||||||
|
tc3 = TC3(self._context)
|
||||||
|
|
||||||
|
mtc = MTC(self._context)
|
||||||
|
|
||||||
|
yObj = {}
|
||||||
|
|
||||||
|
yObj['identifier'] = self.getIdentifier()
|
||||||
|
yObj['variants'] = [self.getVariant(),
|
||||||
|
f"S:{dc3.getVariant()}",
|
||||||
|
f"S:{tc3.getVariant()}",
|
||||||
|
mtc.getVariant()]
|
||||||
|
|
||||||
|
yObj['payload'] = self.getPayload(dc3.getPayload(),
|
||||||
|
tc3.getPayload())
|
||||||
|
|
||||||
|
yObj['assertSelectors'] = ['M', 'SD', 'ST', 'MT']
|
||||||
|
yObj['assertFuncs'] = [self.assertFunc,
|
||||||
|
dc3.createAssertFunc(),
|
||||||
|
tc3.createAssertFunc(),
|
||||||
|
mtc.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj['shouldFail'] = (self.shouldFail()
|
||||||
|
| dc3.shouldFail()
|
||||||
|
| tc3.shouldFail()
|
||||||
|
| mtc.shouldFail())
|
||||||
|
yieldObj = {'target': yObj}
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
dc3_p = DC3(self._context, createPresets = True)
|
||||||
|
tc3_p = TC3(self._context, createPresets = True)
|
||||||
|
|
||||||
|
mtc_p = MTC(self._context, createPresets = True)
|
||||||
|
|
||||||
|
yObj_p = {}
|
||||||
|
|
||||||
|
yObj_p['identifier'] = self.getIdentifier()
|
||||||
|
yObj_p['variants'] = [self.getVariant(),
|
||||||
|
f"S:{dc3_p.getVariant()}",
|
||||||
|
f"S:{tc3_p.getVariant()}",
|
||||||
|
mtc_p.getVariant()]
|
||||||
|
|
||||||
|
yObj_p['payload'] = self.getPayload(dc3_p.getPayload(),
|
||||||
|
tc3_p.getPayload())
|
||||||
|
|
||||||
|
yObj_p['assertSelectors'] = ['M', 'SD', 'ST', 'MT']
|
||||||
|
yObj_p['assertFuncs'] = [self.assertFunc,
|
||||||
|
dc3_p.createAssertFunc(),
|
||||||
|
tc3_p.createAssertFunc(),
|
||||||
|
mtc_p.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj_p['shouldFail'] = (self.shouldFail()
|
||||||
|
| dc3_p.shouldFail()
|
||||||
|
| tc3_p.shouldFail()
|
||||||
|
| mtc_p.shouldFail())
|
||||||
|
yieldObj['preset'] = yObj_p
|
||||||
|
|
||||||
|
yield yieldObj
|
||||||
145
src/ffx/test/media_combinator_4.py
Normal file
145
src/ffx/test/media_combinator_4.py
Normal file
@@ -0,0 +1,145 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .media_combinator import MediaCombinator
|
||||||
|
|
||||||
|
from .disposition_combinator_2 import DispositionCombinator2
|
||||||
|
from .track_tag_combinator_2 import TrackTagCombinator2
|
||||||
|
from .permutation_combinator_2 import PermutationCombinator2
|
||||||
|
from .media_tag_combinator import MediaTagCombinator
|
||||||
|
|
||||||
|
|
||||||
|
class MediaCombinator4(MediaCombinator):
|
||||||
|
|
||||||
|
VARIANT = 'VAA'
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return MediaCombinator4.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self,
|
||||||
|
audioDispositionTuple = (set(), set()),
|
||||||
|
audioTagTuple = ({}, {})):
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.VIDEO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor0 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.AUDIO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = audioDispositionTuple[0]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = audioTagTuple[0]
|
||||||
|
trackDescriptor1 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.AUDIO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = audioDispositionTuple[1]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = audioTagTuple[1]
|
||||||
|
trackDescriptor2 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[MediaDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY] = [trackDescriptor0,
|
||||||
|
trackDescriptor1,
|
||||||
|
trackDescriptor2]
|
||||||
|
|
||||||
|
mediaDescriptor = MediaDescriptor(**kwargs)
|
||||||
|
# mediaDescriptor.reindexSubIndices()
|
||||||
|
|
||||||
|
return mediaDescriptor
|
||||||
|
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
|
||||||
|
for MTC in MediaTagCombinator.getAllClassReferences():
|
||||||
|
for DC2 in DispositionCombinator2.getAllClassReferences():
|
||||||
|
for TC2 in TrackTagCombinator2.getAllClassReferences():
|
||||||
|
|
||||||
|
dc2 = DC2(self._context)
|
||||||
|
tc2 = TC2(self._context)
|
||||||
|
|
||||||
|
mtc = MTC(self._context)
|
||||||
|
|
||||||
|
yObj = {}
|
||||||
|
|
||||||
|
yObj['identifier'] = self.getIdentifier()
|
||||||
|
yObj['variants'] = [self.getVariant(),
|
||||||
|
f"A:{dc2.getVariant()}",
|
||||||
|
f"A:{tc2.getVariant()}",
|
||||||
|
mtc.getVariant()]
|
||||||
|
|
||||||
|
yObj['payload'] = self.getPayload(dc2.getPayload(),
|
||||||
|
tc2.getPayload())
|
||||||
|
|
||||||
|
yObj['assertSelectors'] = ['M', 'AD', 'AT', 'MT']
|
||||||
|
yObj['assertFuncs'] = [self.assertFunc,
|
||||||
|
dc2.createAssertFunc(),
|
||||||
|
tc2.createAssertFunc(),
|
||||||
|
mtc.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj['shouldFail'] = (self.shouldFail()
|
||||||
|
| dc2.shouldFail()
|
||||||
|
| tc2.shouldFail()
|
||||||
|
| mtc.shouldFail())
|
||||||
|
yieldObj = {'target': yObj}
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
dc2_p = DC2(self._context, createPresets = True)
|
||||||
|
tc2_p = TC2(self._context, createPresets = True)
|
||||||
|
|
||||||
|
mtc_p = MTC(self._context, createPresets = True)
|
||||||
|
|
||||||
|
yObj_p = {}
|
||||||
|
|
||||||
|
yObj_p['identifier'] = self.getIdentifier()
|
||||||
|
yObj_p['variants'] = [self.getVariant(),
|
||||||
|
f"A:{dc2_p.getVariant()}",
|
||||||
|
f"A:{tc2_p.getVariant()}",
|
||||||
|
mtc_p.getVariant()]
|
||||||
|
|
||||||
|
yObj_p['payload'] = self.getPayload(dc2_p.getPayload(),
|
||||||
|
tc2_p.getPayload())
|
||||||
|
|
||||||
|
yObj_p['assertSelectors'] = ['M', 'AD', 'AT', 'MT']
|
||||||
|
yObj_p['assertFuncs'] = [self.assertFunc,
|
||||||
|
dc2_p.createAssertFunc(),
|
||||||
|
tc2_p.createAssertFunc(),
|
||||||
|
mtc_p.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj_p['shouldFail'] = (self.shouldFail()
|
||||||
|
| dc2_p.shouldFail()
|
||||||
|
| tc2_p.shouldFail()
|
||||||
|
| mtc_p.shouldFail())
|
||||||
|
yieldObj['preset'] = yObj_p
|
||||||
|
|
||||||
|
yield yieldObj
|
||||||
155
src/ffx/test/media_combinator_5.py
Normal file
155
src/ffx/test/media_combinator_5.py
Normal file
@@ -0,0 +1,155 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .media_combinator import MediaCombinator
|
||||||
|
|
||||||
|
from .disposition_combinator_2 import DispositionCombinator2
|
||||||
|
from .track_tag_combinator_2 import TrackTagCombinator2
|
||||||
|
from .permutation_combinator_2 import PermutationCombinator2
|
||||||
|
from .media_tag_combinator import MediaTagCombinator
|
||||||
|
|
||||||
|
|
||||||
|
class MediaCombinator5(MediaCombinator):
|
||||||
|
|
||||||
|
VARIANT = 'VAAS'
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return MediaCombinator5.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self,
|
||||||
|
audioDispositionTuple = (set(), set()),
|
||||||
|
audioTagTuple = ({}, {})):
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.VIDEO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor0 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.AUDIO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = audioDispositionTuple[0]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = audioTagTuple[0]
|
||||||
|
trackDescriptor1 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.AUDIO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = audioDispositionTuple[1]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = audioTagTuple[1]
|
||||||
|
trackDescriptor2 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 3
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 3
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.SUBTITLE
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor3 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[MediaDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY] = [trackDescriptor0,
|
||||||
|
trackDescriptor1,
|
||||||
|
trackDescriptor2,
|
||||||
|
trackDescriptor3]
|
||||||
|
|
||||||
|
mediaDescriptor = MediaDescriptor(**kwargs)
|
||||||
|
# mediaDescriptor.reindexSubIndices()
|
||||||
|
|
||||||
|
return mediaDescriptor
|
||||||
|
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
|
||||||
|
for MTC in MediaTagCombinator.getAllClassReferences():
|
||||||
|
for DC2 in DispositionCombinator2.getAllClassReferences():
|
||||||
|
for TC2 in TrackTagCombinator2.getAllClassReferences():
|
||||||
|
|
||||||
|
dc2 = DC2(self._context)
|
||||||
|
tc2 = TC2(self._context)
|
||||||
|
|
||||||
|
mtc = MTC(self._context)
|
||||||
|
|
||||||
|
yObj = {}
|
||||||
|
|
||||||
|
yObj['identifier'] = self.getIdentifier()
|
||||||
|
yObj['variants'] = [self.getVariant(),
|
||||||
|
f"A:{dc2.getVariant()}",
|
||||||
|
f"A:{tc2.getVariant()}",
|
||||||
|
mtc.getVariant()]
|
||||||
|
|
||||||
|
yObj['payload'] = self.getPayload(dc2.getPayload(),
|
||||||
|
tc2.getPayload())
|
||||||
|
|
||||||
|
yObj['assertSelectors'] = ['M', 'AD', 'AT', 'MT']
|
||||||
|
yObj['assertFuncs'] = [self.assertFunc,
|
||||||
|
dc2.createAssertFunc(),
|
||||||
|
tc2.createAssertFunc(),
|
||||||
|
mtc.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj['shouldFail'] = (self.shouldFail()
|
||||||
|
| dc2.shouldFail()
|
||||||
|
| tc2.shouldFail()
|
||||||
|
| mtc.shouldFail())
|
||||||
|
yieldObj = {'target': yObj}
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
dc2_p = DC2(self._context, createPresets = True)
|
||||||
|
tc2_p = TC2(self._context, createPresets = True)
|
||||||
|
|
||||||
|
mtc_p = MTC(self._context, createPresets = True)
|
||||||
|
|
||||||
|
yObj_p = {}
|
||||||
|
|
||||||
|
yObj_p['identifier'] = self.getIdentifier()
|
||||||
|
yObj_p['variants'] = [self.getVariant(),
|
||||||
|
f"A:{dc2_p.getVariant()}",
|
||||||
|
f"A:{tc2_p.getVariant()}",
|
||||||
|
mtc_p.getVariant()]
|
||||||
|
|
||||||
|
yObj_p['payload'] = self.getPayload(dc2_p.getPayload(),
|
||||||
|
tc2_p.getPayload())
|
||||||
|
|
||||||
|
yObj_p['assertSelectors'] = ['M', 'AD', 'AT', 'MT']
|
||||||
|
yObj_p['assertFuncs'] = [self.assertFunc,
|
||||||
|
dc2_p.createAssertFunc(),
|
||||||
|
tc2_p.createAssertFunc(),
|
||||||
|
mtc_p.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj_p['shouldFail'] = (self.shouldFail()
|
||||||
|
| dc2_p.shouldFail()
|
||||||
|
| tc2_p.shouldFail()
|
||||||
|
| mtc_p.shouldFail())
|
||||||
|
yieldObj['preset'] = yObj_p
|
||||||
|
|
||||||
|
yield yieldObj
|
||||||
191
src/ffx/test/media_combinator_6.py
Normal file
191
src/ffx/test/media_combinator_6.py
Normal file
@@ -0,0 +1,191 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .media_combinator import MediaCombinator
|
||||||
|
|
||||||
|
from .disposition_combinator_2 import DispositionCombinator2
|
||||||
|
from .track_tag_combinator_2 import TrackTagCombinator2
|
||||||
|
from .permutation_combinator_2 import PermutationCombinator2
|
||||||
|
from .media_tag_combinator import MediaTagCombinator
|
||||||
|
|
||||||
|
|
||||||
|
class MediaCombinator6(MediaCombinator):
|
||||||
|
|
||||||
|
VARIANT = 'VAASS'
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return MediaCombinator6.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self,
|
||||||
|
audioDispositionTuple = (set(), set()),
|
||||||
|
audioTagTuple = ({}, {}),
|
||||||
|
subtitleDispositionTuple = (set(), set()),
|
||||||
|
subtitleTagTuple = ({}, {})):
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.VIDEO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor0 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.AUDIO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = audioDispositionTuple[0]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = audioTagTuple[0]
|
||||||
|
trackDescriptor1 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.AUDIO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = audioDispositionTuple[1]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = audioTagTuple[1]
|
||||||
|
trackDescriptor2 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 3
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 3
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.SUBTITLE
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = subtitleDispositionTuple[0]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = subtitleTagTuple[0]
|
||||||
|
trackDescriptor3 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 4
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 4
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.SUBTITLE
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = subtitleDispositionTuple[1]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = subtitleTagTuple[1]
|
||||||
|
trackDescriptor4 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[MediaDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY] = [trackDescriptor0,
|
||||||
|
trackDescriptor1,
|
||||||
|
trackDescriptor2,
|
||||||
|
trackDescriptor3,
|
||||||
|
trackDescriptor4]
|
||||||
|
|
||||||
|
mediaDescriptor = MediaDescriptor(**kwargs)
|
||||||
|
# mediaDescriptor.reindexSubIndices()
|
||||||
|
|
||||||
|
return mediaDescriptor
|
||||||
|
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
|
||||||
|
for MTC in MediaTagCombinator.getAllClassReferences():
|
||||||
|
for DC2_A in DispositionCombinator2.getAllClassReferences():
|
||||||
|
for TC2_A in TrackTagCombinator2.getAllClassReferences():
|
||||||
|
for DC2_S in DispositionCombinator2.getAllClassReferences():
|
||||||
|
for TC2_S in TrackTagCombinator2.getAllClassReferences():
|
||||||
|
|
||||||
|
dc2a = DC2_A(self._context)
|
||||||
|
tc2a = TC2_A(self._context)
|
||||||
|
dc2s = DC2_S(self._context)
|
||||||
|
tc2s = TC2_S(self._context)
|
||||||
|
|
||||||
|
mtc = MTC(self._context)
|
||||||
|
|
||||||
|
yObj = {}
|
||||||
|
|
||||||
|
yObj['identifier'] = self.getIdentifier()
|
||||||
|
yObj['variants'] = [self.getVariant(),
|
||||||
|
f"A:{dc2a.getVariant()}",
|
||||||
|
f"A:{tc2a.getVariant()}",
|
||||||
|
f"S:{dc2s.getVariant()}",
|
||||||
|
f"S:{tc2s.getVariant()}",
|
||||||
|
mtc.getVariant()]
|
||||||
|
|
||||||
|
yObj['payload'] = self.getPayload(dc2a.getPayload(),
|
||||||
|
tc2a.getPayload(),
|
||||||
|
dc2s.getPayload(),
|
||||||
|
tc2s.getPayload())
|
||||||
|
|
||||||
|
yObj['assertSelectors'] = ['M', 'AD', 'AT', 'SD', 'ST', 'MT']
|
||||||
|
yObj['assertFuncs'] = [self.assertFunc,
|
||||||
|
dc2a.createAssertFunc(),
|
||||||
|
tc2a.createAssertFunc(),
|
||||||
|
dc2s.createAssertFunc(),
|
||||||
|
tc2s.createAssertFunc(),
|
||||||
|
mtc.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj['shouldFail'] = (self.shouldFail()
|
||||||
|
| dc2a.shouldFail()
|
||||||
|
| tc2a.shouldFail()
|
||||||
|
| dc2s.shouldFail()
|
||||||
|
| tc2s.shouldFail()
|
||||||
|
| mtc.shouldFail())
|
||||||
|
yieldObj = {'target': yObj}
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
dc2a_p = DC2_A(self._context, createPresets = True)
|
||||||
|
tc2a_p = TC2_A(self._context, createPresets = True)
|
||||||
|
dc2s_p = DC2_S(self._context, createPresets = True)
|
||||||
|
tc2s_p = TC2_S(self._context, createPresets = True)
|
||||||
|
|
||||||
|
mtc_p = MTC(self._context, createPresets = True)
|
||||||
|
|
||||||
|
yObj_p = {}
|
||||||
|
|
||||||
|
yObj_p['identifier'] = self.getIdentifier()
|
||||||
|
yObj_p['variants'] = [self.getVariant(),
|
||||||
|
f"A:{dc2a_p.getVariant()}",
|
||||||
|
f"A:{tc2a_p.getVariant()}",
|
||||||
|
f"S:{dc2s_p.getVariant()}",
|
||||||
|
f"S:{tc2s_p.getVariant()}",
|
||||||
|
mtc_p.getVariant()]
|
||||||
|
|
||||||
|
yObj_p['payload'] = self.getPayload(dc2a_p.getPayload(),
|
||||||
|
tc2a_p.getPayload(),
|
||||||
|
dc2s_p.getPayload(),
|
||||||
|
tc2s_p.getPayload())
|
||||||
|
|
||||||
|
yObj_p['assertSelectors'] = ['M', 'AD', 'AT', 'SD', 'ST', 'MT']
|
||||||
|
yObj_p['assertFuncs'] = [self.assertFunc,
|
||||||
|
dc2a_p.createAssertFunc(),
|
||||||
|
tc2a_p.createAssertFunc(),
|
||||||
|
dc2s_p.createAssertFunc(),
|
||||||
|
tc2s_p.createAssertFunc(),
|
||||||
|
mtc_p.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj_p['shouldFail'] = (self.shouldFail()
|
||||||
|
| dc2a_p.shouldFail()
|
||||||
|
| tc2a_p.shouldFail()
|
||||||
|
| dc2s_p.shouldFail()
|
||||||
|
| tc2s_p.shouldFail()
|
||||||
|
| mtc_p.shouldFail())
|
||||||
|
yieldObj['preset'] = yObj_p
|
||||||
|
|
||||||
|
yield yieldObj
|
||||||
230
src/ffx/test/media_combinator_7.py
Normal file
230
src/ffx/test/media_combinator_7.py
Normal file
@@ -0,0 +1,230 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .media_combinator import MediaCombinator
|
||||||
|
|
||||||
|
from .disposition_combinator_2 import DispositionCombinator2
|
||||||
|
from .disposition_combinator_3 import DispositionCombinator3
|
||||||
|
from .track_tag_combinator_2 import TrackTagCombinator2
|
||||||
|
from .track_tag_combinator_3 import TrackTagCombinator3
|
||||||
|
from .permutation_combinator_2 import PermutationCombinator2
|
||||||
|
from .permutation_combinator_3 import PermutationCombinator3
|
||||||
|
from .media_tag_combinator import MediaTagCombinator
|
||||||
|
|
||||||
|
class MediaCombinator7(MediaCombinator):
|
||||||
|
|
||||||
|
VARIANT = 'VAASSS'
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return MediaCombinator7.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self,
|
||||||
|
audioPermutation,
|
||||||
|
subtitlePermutation,
|
||||||
|
audioDispositionTuple = (set(), set()),
|
||||||
|
audioTagTuple = ({}, {}),
|
||||||
|
subtitleDispositionTuple = (set(), set(), set()),
|
||||||
|
subtitleTagTuple = ({}, {}, {})):
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.VIDEO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
trackDescriptor0 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.AUDIO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = audioDispositionTuple[0]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = audioTagTuple[0]
|
||||||
|
trackDescriptor1 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.AUDIO
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = audioDispositionTuple[1]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = audioTagTuple[1]
|
||||||
|
trackDescriptor2 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 3
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 3
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.SUBTITLE
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 0
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = subtitleDispositionTuple[0]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = subtitleTagTuple[0]
|
||||||
|
trackDescriptor3 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 4
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 4
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.SUBTITLE
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 1
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = subtitleDispositionTuple[1]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = subtitleTagTuple[1]
|
||||||
|
trackDescriptor4 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[TrackDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[TrackDescriptor.INDEX_KEY] = 5
|
||||||
|
kwargs[TrackDescriptor.SOURCE_INDEX_KEY] = 5
|
||||||
|
kwargs[TrackDescriptor.TRACK_TYPE_KEY] = TrackType.SUBTITLE
|
||||||
|
kwargs[TrackDescriptor.SUB_INDEX_KEY] = 2
|
||||||
|
kwargs[TrackDescriptor.DISPOSITION_SET_KEY] = subtitleDispositionTuple[2]
|
||||||
|
kwargs[TrackDescriptor.TAGS_KEY] = subtitleTagTuple[2]
|
||||||
|
trackDescriptor5 = TrackDescriptor(**kwargs)
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[MediaDescriptor.CONTEXT_KEY] = self._context
|
||||||
|
kwargs[MediaDescriptor.TRACK_DESCRIPTOR_LIST_KEY] = [trackDescriptor0,
|
||||||
|
trackDescriptor1,
|
||||||
|
trackDescriptor2,
|
||||||
|
trackDescriptor3,
|
||||||
|
trackDescriptor4,
|
||||||
|
trackDescriptor5]
|
||||||
|
|
||||||
|
mediaDescriptor = MediaDescriptor(**kwargs)
|
||||||
|
# mediaDescriptor.reindexSubIndices()
|
||||||
|
|
||||||
|
return mediaDescriptor
|
||||||
|
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
|
||||||
|
|
||||||
|
pc2 = PermutationCombinator2(self._context)
|
||||||
|
pc3 = PermutationCombinator3(self._context)
|
||||||
|
|
||||||
|
for MTC in MediaTagCombinator.getAllClassReferences():
|
||||||
|
for DC2_A in DispositionCombinator2.getAllClassReferences():
|
||||||
|
for TC2_A in TrackTagCombinator2.getAllClassReferences():
|
||||||
|
for DC3_S in DispositionCombinator3.getAllClassReferences():
|
||||||
|
for TC3_S in TrackTagCombinator3.getAllClassReferences():
|
||||||
|
for p2y in pc2.getYield():
|
||||||
|
for p3y in pc3.getYield():
|
||||||
|
|
||||||
|
dc2a = DC2_A(self._context)
|
||||||
|
tc2a = TC2_A(self._context)
|
||||||
|
dc3s = DC3_S(self._context)
|
||||||
|
tc3s = TC3_S(self._context)
|
||||||
|
|
||||||
|
mtc = MTC(self._context)
|
||||||
|
|
||||||
|
yObj = {}
|
||||||
|
|
||||||
|
yObj['identifier'] = self.getIdentifier()
|
||||||
|
yObj['variants'] = [self.getVariant(),
|
||||||
|
f"A:{p2y['variant']}",
|
||||||
|
f"S:{p3y['variant']}",
|
||||||
|
f"A:{dc2a.getVariant()}",
|
||||||
|
f"A:{tc2a.getVariant()}",
|
||||||
|
f"S:{dc3s.getVariant()}",
|
||||||
|
f"S:{tc3s.getVariant()}",
|
||||||
|
mtc.getVariant()]
|
||||||
|
|
||||||
|
yObj['payload'] = self.getPayload(p2y['permutation'],
|
||||||
|
p3y['permutation'],
|
||||||
|
dc2a.getPayload(),
|
||||||
|
tc2a.getPayload(),
|
||||||
|
dc3s.getPayload(),
|
||||||
|
tc3s.getPayload())
|
||||||
|
|
||||||
|
yObj['assertSelectors'] = ['M', 'AP', 'SP', 'AD', 'AT', 'SD', 'ST', 'MT']
|
||||||
|
|
||||||
|
yObj['assertFuncs'] = [self.assertFunc,
|
||||||
|
p2y['assertFunc'],
|
||||||
|
p3y['assertFunc'],
|
||||||
|
dc2a.createAssertFunc(),
|
||||||
|
tc2a.createAssertFunc(),
|
||||||
|
dc3s.createAssertFunc(),
|
||||||
|
tc3s.createAssertFunc(),
|
||||||
|
mtc.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj['shouldFail'] = (self.shouldFail()
|
||||||
|
| p2y['shouldFail']
|
||||||
|
| p3y['shouldFail']
|
||||||
|
| dc2a.shouldFail()
|
||||||
|
| tc2a.shouldFail()
|
||||||
|
| dc3s.shouldFail()
|
||||||
|
| tc3s.shouldFail()
|
||||||
|
| mtc.shouldFail())
|
||||||
|
yieldObj = {'target': yObj}
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
dc2a_p = DC2_A(self._context, createPresets = True)
|
||||||
|
tc2a_p = TC2_A(self._context, createPresets = True)
|
||||||
|
dc3s_p = DC3_S(self._context, createPresets = True)
|
||||||
|
tc3s_p = TC3_S(self._context, createPresets = True)
|
||||||
|
|
||||||
|
mtc_p = MTC(self._context, createPresets = True)
|
||||||
|
|
||||||
|
yObj_p = {}
|
||||||
|
|
||||||
|
yObj_p['identifier'] = self.getIdentifier()
|
||||||
|
yObj_p['variants'] = [self.getVariant(),
|
||||||
|
f"A:{p2y['variant']}",
|
||||||
|
f"S:{p3y['variant']}",
|
||||||
|
f"A:{dc2a_p.getVariant()}",
|
||||||
|
f"A:{tc2a_p.getVariant()}",
|
||||||
|
f"S:{dc3s_p.getVariant()}",
|
||||||
|
f"S:{tc3s_p.getVariant()}",
|
||||||
|
mtc_p.getVariant()]
|
||||||
|
|
||||||
|
yObj_p['payload'] = self.getPayload(p2y['permutation'],
|
||||||
|
p3y['permutation'],
|
||||||
|
dc2a_p.getPayload(),
|
||||||
|
tc2a_p.getPayload(),
|
||||||
|
dc3s_p.getPayload(),
|
||||||
|
tc3s_p.getPayload())
|
||||||
|
|
||||||
|
yObj_p['assertSelectors'] = ['M', 'AP', 'SP', 'AD', 'AT', 'SD', 'ST', 'MT']
|
||||||
|
|
||||||
|
yObj_p['assertFuncs'] = [self.assertFunc,
|
||||||
|
p2y['assertFunc'],
|
||||||
|
p3y['assertFunc'],
|
||||||
|
dc2a_p.createAssertFunc(),
|
||||||
|
tc2a_p.createAssertFunc(),
|
||||||
|
dc3s_p.createAssertFunc(),
|
||||||
|
tc3s_p.createAssertFunc(),
|
||||||
|
mtc_p.createAssertFunc()]
|
||||||
|
|
||||||
|
yObj_p['shouldFail'] = (self.shouldFail()
|
||||||
|
| p2y['shouldFail']
|
||||||
|
| p3y['shouldFail']
|
||||||
|
| dc2a_p.shouldFail()
|
||||||
|
| tc2a_p.shouldFail()
|
||||||
|
| dc3s_p.shouldFail()
|
||||||
|
| tc3s_p.shouldFail()
|
||||||
|
| mtc_p.shouldFail())
|
||||||
|
yieldObj['preset'] = yObj_p
|
||||||
|
|
||||||
|
yield yieldObj
|
||||||
33
src/ffx/test/media_tag_combinator.py
Normal file
33
src/ffx/test/media_tag_combinator.py
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
class MediaTagCombinator():
|
||||||
|
|
||||||
|
IDENTIFIER = 'mediaTag'
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getIdentifier(self):
|
||||||
|
return MediaTagCombinator.IDENTIFIER
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def list():
|
||||||
|
basePath = os.path.dirname(__file__)
|
||||||
|
return [os.path.basename(p)[21:-3]
|
||||||
|
for p
|
||||||
|
in glob.glob(f"{ basePath }/media_tag_combinator_*.py", recursive = True)
|
||||||
|
if p != __file__]
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getClassReference(identifier):
|
||||||
|
importlib.import_module(f"ffx.test.media_tag_combinator_{ identifier }")
|
||||||
|
for name, obj in inspect.getmembers(sys.modules[f"ffx.test.media_tag_combinator_{ identifier }"]):
|
||||||
|
#HINT: Excluding MediaCombinator as it seems to be included by import (?)
|
||||||
|
if inspect.isclass(obj) and name != 'MediaTagCombinator' and name.startswith('MediaTagCombinator'):
|
||||||
|
return obj
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getAllClassReferences():
|
||||||
|
return [MediaTagCombinator.getClassReference(i) for i in MediaTagCombinator.list()]
|
||||||
55
src/ffx/test/media_tag_combinator_0.py
Normal file
55
src/ffx/test/media_tag_combinator_0.py
Normal file
@@ -0,0 +1,55 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .media_tag_combinator import MediaTagCombinator
|
||||||
|
|
||||||
|
|
||||||
|
class MediaTagCombinator0(MediaTagCombinator):
|
||||||
|
|
||||||
|
VARIANT = 'MT0'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return MediaTagCombinator0.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
mediaTags = {}
|
||||||
|
if self.__createPresets:
|
||||||
|
mediaTags['THIS_IS'] = 'FFX'
|
||||||
|
return mediaTags
|
||||||
|
|
||||||
|
|
||||||
|
def createAssertFunc(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
def f(assertObj: dict = {}):
|
||||||
|
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tags'")
|
||||||
|
mediaTags = assertObj['tags']
|
||||||
|
|
||||||
|
assert ('THIS_IS' in mediaTags.keys() and mediaTags.keys()['THIS_IS'] == 'FFX'
|
||||||
|
), "Media tag 'THIS_IS' was not preserved"
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
def f(assertObj: dict = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
return f
|
||||||
|
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
59
src/ffx/test/media_tag_combinator_1.py
Normal file
59
src/ffx/test/media_tag_combinator_1.py
Normal file
@@ -0,0 +1,59 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .media_tag_combinator import MediaTagCombinator
|
||||||
|
|
||||||
|
class MediaTagCombinator1(MediaTagCombinator):
|
||||||
|
|
||||||
|
VARIANT = 'MT1'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return MediaTagCombinator1.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
mediaTags = {'From': 'Encoders'}
|
||||||
|
if self.__createPresets:
|
||||||
|
mediaTags['THIS_IS'] = 'FFX'
|
||||||
|
return mediaTags
|
||||||
|
|
||||||
|
|
||||||
|
def createAssertFunc(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
def f(assertObj: dict = {}):
|
||||||
|
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tags'")
|
||||||
|
mediaTags = assertObj['tags']
|
||||||
|
|
||||||
|
assert ('From' in mediaTags.keys()
|
||||||
|
), "'From' not in media tag keys"
|
||||||
|
assert (mediaTags.keys()['From'] == 'Encoders'
|
||||||
|
), "Media tag value not 'Encoders' for key 'To'"
|
||||||
|
|
||||||
|
assert ('THIS_IS' in mediaTags.keys() and mediaTags.keys()['THIS_IS'] == 'FFX'
|
||||||
|
), "Media tag 'THIS_IS' was not preserved"
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
def f(assertObj: dict = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
return f
|
||||||
|
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
64
src/ffx/test/media_tag_combinator_2.py
Normal file
64
src/ffx/test/media_tag_combinator_2.py
Normal file
@@ -0,0 +1,64 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
|
||||||
|
from .media_tag_combinator import MediaTagCombinator
|
||||||
|
|
||||||
|
class MediaTagCombinator2(MediaTagCombinator):
|
||||||
|
|
||||||
|
VARIANT = 'MT2'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return MediaTagCombinator2.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
mediaTags = {'To': 'Fanz',
|
||||||
|
'Yolo': 'Holo'}
|
||||||
|
if self.__createPresets:
|
||||||
|
mediaTags['THIS_IS'] = 'FFX'
|
||||||
|
return mediaTags
|
||||||
|
|
||||||
|
|
||||||
|
def createAssertFunc(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
def f(assertObj: dict = {}):
|
||||||
|
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tags'")
|
||||||
|
mediaTags = assertObj['tags']
|
||||||
|
|
||||||
|
assert ('To' in mediaTags.keys()
|
||||||
|
), "'To' not in media tag keys"
|
||||||
|
assert (mediaTags.keys()['To'] == 'Fanz'
|
||||||
|
), "Media tag value not 'Fanz' for key 'To'"
|
||||||
|
assert ('Yolo' in mediaTags.keys()
|
||||||
|
), "'Yolo' not in media tag keys"
|
||||||
|
assert (mediaTags.keys()['Yolo'] == 'Holo'
|
||||||
|
), "Media tag value not 'Holo' for key 'Yolo'"
|
||||||
|
|
||||||
|
assert ('THIS_IS' in mediaTags.keys() and mediaTags.keys()['THIS_IS'] == 'FFX'
|
||||||
|
), "Media tag 'THIS_IS' was not preserved"
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
def f(assertObj: dict = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
return f
|
||||||
|
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
38
src/ffx/test/permutation_combinator_2.py
Normal file
38
src/ffx/test/permutation_combinator_2.py
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
class PermutationCombinator2():
|
||||||
|
|
||||||
|
IDENTIFIER = 'permutation2'
|
||||||
|
|
||||||
|
PERMUTATION_LIST = [
|
||||||
|
[0,1],
|
||||||
|
[1,0]
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getIdentifier(self):
|
||||||
|
return PermutationCombinator2.IDENTIFIER
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self, permutationIndex):
|
||||||
|
return {
|
||||||
|
'variant': f"P{permutationIndex}",
|
||||||
|
'permutation': PermutationCombinator2.PERMUTATION_LIST[permutationIndex],
|
||||||
|
'assertFunc': self.createAssertFunc(),
|
||||||
|
'shouldFail': self.shouldFail()
|
||||||
|
}
|
||||||
|
|
||||||
|
def createAssertFunc(self):
|
||||||
|
def f(testObj = {}):
|
||||||
|
pass
|
||||||
|
return f
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
for permutationIndex in range(len(PermutationCombinator2.PERMUTATION_LIST)):
|
||||||
|
yield self.getPayload(permutationIndex)
|
||||||
39
src/ffx/test/permutation_combinator_3.py
Normal file
39
src/ffx/test/permutation_combinator_3.py
Normal file
@@ -0,0 +1,39 @@
|
|||||||
|
class PermutationCombinator3():
|
||||||
|
|
||||||
|
IDENTIFIER = 'permutation3'
|
||||||
|
|
||||||
|
PERMUTATION_LIST = [
|
||||||
|
[0,1,2],
|
||||||
|
[0,2,1],
|
||||||
|
[1,2,0]
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getIdentifier(self):
|
||||||
|
return PermutationCombinator3.IDENTIFIER
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self, permutationIndex):
|
||||||
|
return {
|
||||||
|
'variant': f"P{permutationIndex}",
|
||||||
|
'permutation': PermutationCombinator3.PERMUTATION_LIST[permutationIndex],
|
||||||
|
'assertFunc': self.createAssertFunc(),
|
||||||
|
'shouldFail': self.shouldFail()
|
||||||
|
}
|
||||||
|
|
||||||
|
def createAssertFunc(self):
|
||||||
|
def f(testObj = {}):
|
||||||
|
pass
|
||||||
|
return f
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
for permutationIndex in range(len(PermutationCombinator3.PERMUTATION_LIST)):
|
||||||
|
yield self.getPayload(permutationIndex)
|
||||||
37
src/ffx/test/release_combinator.py
Normal file
37
src/ffx/test/release_combinator.py
Normal file
@@ -0,0 +1,37 @@
|
|||||||
|
class ReleaseCombinator():
|
||||||
|
|
||||||
|
IDENTIFIER = 'release'
|
||||||
|
|
||||||
|
RELEASE_LIST = [
|
||||||
|
".GerEngSub.AAC.1080pINDICATOR.WebDL.x264-Tanuki",
|
||||||
|
".German.AC3.DL.1080pINDICATOR.BluRay.x264-AST4u",
|
||||||
|
"-720pINDICATOR"
|
||||||
|
]
|
||||||
|
|
||||||
|
def __init__(self, context = None, indicator = ''):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
self.__indicator = indicator
|
||||||
|
|
||||||
|
def getIdentifier(self):
|
||||||
|
return ReleaseCombinator.IDENTIFIER
|
||||||
|
|
||||||
|
def getPayload(self, releaseIndex):
|
||||||
|
releaseStr: str = ReleaseCombinator.RELEASE_LIST[releaseIndex]
|
||||||
|
return {
|
||||||
|
'variant': f"R{releaseIndex}",
|
||||||
|
'release': releaseStr.replace('INDICATOR', f".{self.__indicator}")
|
||||||
|
if self.__indicator else releaseStr.replace('INDICATOR', '')
|
||||||
|
}
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
for releaseIndex in range(len(ReleaseCombinator.RELEASE_LIST)):
|
||||||
|
yield self.getPayload(releaseIndex)
|
||||||
153
src/ffx/test/scenario.py
Normal file
153
src/ffx/test/scenario.py
Normal file
@@ -0,0 +1,153 @@
|
|||||||
|
import os, glob, sys, importlib, glob, inspect
|
||||||
|
|
||||||
|
from ffx.show_controller import ShowController
|
||||||
|
from ffx.pattern_controller import PatternController
|
||||||
|
from ffx.media_controller import MediaController
|
||||||
|
|
||||||
|
from ffx.test.helper import createEmptyDirectory
|
||||||
|
from ffx.database import databaseContext
|
||||||
|
|
||||||
|
class Scenario():
|
||||||
|
"""Scenarios
|
||||||
|
|
||||||
|
Scenario1: MediaTags, Stream-Kombinationen, Dispositions und StreamTags per Kombinatoren
|
||||||
|
|
||||||
|
Scenario2: <pattern> mit 3 Files x Scenario1
|
||||||
|
|
||||||
|
Scenario3: <tmdb+pattern> mit 3 Files (wenn TMDB API Key verfügbar)
|
||||||
|
|
||||||
|
Naming:
|
||||||
|
|
||||||
|
1: test.mkv no tmdb, no pattern
|
||||||
|
2: test_s01e02.mkv
|
||||||
|
|
||||||
|
Operationen:
|
||||||
|
|
||||||
|
tmdb lookup: Set Showname as prefix, append episode name
|
||||||
|
pattern lookup: Set/update tags/dispositions; Filter/Reorder Tracks
|
||||||
|
|
||||||
|
MediaTag-Kombinationen (2)
|
||||||
|
|
||||||
|
0: nichs
|
||||||
|
1: Yolo=Holo
|
||||||
|
|
||||||
|
Stream-Kombinationen (8)
|
||||||
|
|
||||||
|
VA D=1 T=1 =1
|
||||||
|
VAS D=1 T=1 =1
|
||||||
|
VASS D=4 T=4 =16
|
||||||
|
VASSS D=5 T=5 =25
|
||||||
|
VAA D=4 T=4 =16
|
||||||
|
VAAS D=4 T=4 =16
|
||||||
|
VAASS D=16 T=16 =256
|
||||||
|
VAASSS D=20 T=20 =400
|
||||||
|
=731
|
||||||
|
|
||||||
|
Dispositions-Kombinationen (per TrackType)
|
||||||
|
|
||||||
|
0 = keine
|
||||||
|
1 = DEFAULT
|
||||||
|
|
||||||
|
2 Streams (4):
|
||||||
|
|
||||||
|
D1: 00
|
||||||
|
D2: 01
|
||||||
|
D3: 10
|
||||||
|
D4: 11
|
||||||
|
|
||||||
|
3 Streams (5):
|
||||||
|
|
||||||
|
D5: 000
|
||||||
|
D6: 001
|
||||||
|
D7: 010
|
||||||
|
D8: 100
|
||||||
|
D9: 101
|
||||||
|
|
||||||
|
Stream-Tag-Kombinationen (per TrackType)
|
||||||
|
|
||||||
|
0 = keine
|
||||||
|
1 = lang+title
|
||||||
|
|
||||||
|
2 Streams:
|
||||||
|
|
||||||
|
00
|
||||||
|
01
|
||||||
|
10
|
||||||
|
11
|
||||||
|
|
||||||
|
3 Streams:
|
||||||
|
|
||||||
|
000
|
||||||
|
001
|
||||||
|
010
|
||||||
|
100
|
||||||
|
101
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._testDirectory = createEmptyDirectory()
|
||||||
|
self._ffxExecutablePath = os.path.join(
|
||||||
|
os.path.dirname(
|
||||||
|
os.path.dirname(
|
||||||
|
os.path.dirname(__file__))),
|
||||||
|
'ffx.py')
|
||||||
|
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
self._testDbFilePath = os.path.join(self._testDirectory, 'test.db')
|
||||||
|
self.createEmptyTestDatabase()
|
||||||
|
|
||||||
|
# Convenience
|
||||||
|
self._niceness = self._context['resource_limits']['niceness'] if 'resource_limits' in self._context.keys() and 'niceness' in self._context['resource_limits'].keys() else 99
|
||||||
|
self._cpuPercent = self._context['resource_limits']['cpu_percent'] if 'resource_limits' in self._context.keys() and 'cpu_percent' in self._context['resource_limits'].keys() else 99
|
||||||
|
|
||||||
|
|
||||||
|
def createEmptyTestDatabase(self):
|
||||||
|
|
||||||
|
if not self._context['database'] is None:
|
||||||
|
self._context['database']['engine'].dispose()
|
||||||
|
|
||||||
|
if os.path.isfile(self._testDbFilePath):
|
||||||
|
os.unlink(self._testDbFilePath)
|
||||||
|
self._context['database'] = None
|
||||||
|
|
||||||
|
self._logger.debug(f"Creating test db with path {self._testDbFilePath}")
|
||||||
|
self._context['database'] = databaseContext(databasePath=self._testDbFilePath)
|
||||||
|
|
||||||
|
self._sc = ShowController(context = self._context)
|
||||||
|
self._pc = PatternController(context = self._context)
|
||||||
|
self._mc = MediaController(context = self._context)
|
||||||
|
|
||||||
|
|
||||||
|
def clearTestDirectory(self):
|
||||||
|
testFiles = glob.glob(f"{self._testDirectory}/*")
|
||||||
|
for f in testFiles:
|
||||||
|
os.remove(f)
|
||||||
|
|
||||||
|
def getFilePathsInTestDirectory(self):
|
||||||
|
return [f for f in glob.glob(f"{self._testDirectory}/*")]
|
||||||
|
|
||||||
|
def getFilenamesInTestDirectory(self):
|
||||||
|
return [os.path.basename(f) for f in self.getFilePathsInTestDirectory()]
|
||||||
|
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def list():
|
||||||
|
basePath = os.path.dirname(__file__)
|
||||||
|
return [os.path.basename(p)[9:-3]
|
||||||
|
for p
|
||||||
|
in glob.glob(f"{ basePath }/scenario_*.py", recursive = True)
|
||||||
|
if p != __file__]
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getClassReference(identifier):
|
||||||
|
importlib.import_module(f"ffx.test.scenario_{ identifier }")
|
||||||
|
for name, obj in inspect.getmembers(sys.modules[f"ffx.test.scenario_{ identifier }"]):
|
||||||
|
#HINT: Excluding Scenario as it seems to be included by import (?)
|
||||||
|
if inspect.isclass(obj) and name != 'Scenario' and name.startswith('Scenario'):
|
||||||
|
return obj
|
||||||
173
src/ffx/test/scenario_1.py
Normal file
173
src/ffx/test/scenario_1.py
Normal file
@@ -0,0 +1,173 @@
|
|||||||
|
import os, sys, click, glob
|
||||||
|
|
||||||
|
from .scenario import Scenario
|
||||||
|
|
||||||
|
from ffx.test.helper import createMediaTestFile
|
||||||
|
from ffx.process import executeProcess
|
||||||
|
|
||||||
|
from ffx.file_properties import FileProperties
|
||||||
|
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
|
||||||
|
from ffx.test.media_combinator_0 import MediaCombinator0
|
||||||
|
|
||||||
|
from ffx.test.basename_combinator import BasenameCombinator
|
||||||
|
|
||||||
|
|
||||||
|
class Scenario1(Scenario):
|
||||||
|
"""Creating file VAa, h264/aac/aac
|
||||||
|
Converting to VaA, vp9/opus/opus
|
||||||
|
No tmdb, default parameters"""
|
||||||
|
|
||||||
|
TEST_FILE_EXTENSION = 'mkv'
|
||||||
|
EXPECTED_FILE_EXTENSION = 'webm'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context):
|
||||||
|
|
||||||
|
context['use_tmdb'] = False
|
||||||
|
context['use_pattern'] = False
|
||||||
|
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
def getScenario(self):
|
||||||
|
return self.__class__.__name__[8:]
|
||||||
|
|
||||||
|
|
||||||
|
def job(self, yieldObj: dict):
|
||||||
|
|
||||||
|
testContext = self._context.copy()
|
||||||
|
|
||||||
|
identifier = yieldObj['identifier']
|
||||||
|
variantList = yieldObj['variants']
|
||||||
|
|
||||||
|
variantIdentifier = '-'.join(variantList)
|
||||||
|
variantLabel = f"{self.__class__.__name__} Variant {variantIdentifier}"
|
||||||
|
|
||||||
|
mc0 = MediaCombinator0(context = testContext)
|
||||||
|
sourceMediaDescriptor: MediaDescriptor = mc0.getPayload()
|
||||||
|
|
||||||
|
assertSelectorList: list = yieldObj['assertSelectors']
|
||||||
|
assertFuncList = yieldObj['assertFuncs']
|
||||||
|
shouldFail = yieldObj['shouldFail']
|
||||||
|
|
||||||
|
variantPayload = yieldObj['payload']
|
||||||
|
variantBasename = variantPayload['basename']
|
||||||
|
variantFilenameLabel = variantPayload['label']
|
||||||
|
expectedBasename = variantPayload['expectedBasename']
|
||||||
|
|
||||||
|
variantFilename = f"{variantBasename}.{Scenario1.TEST_FILE_EXTENSION}"
|
||||||
|
expectedFilename = f"{expectedBasename}.{Scenario1.EXPECTED_FILE_EXTENSION}"
|
||||||
|
|
||||||
|
|
||||||
|
if self._context['test_variant'] and not variantIdentifier.startswith(self._context['test_variant']):
|
||||||
|
return
|
||||||
|
|
||||||
|
if (self._context['test_limit'] and (self._context['test_passed_counter'] + self._context['test_failed_counter'])
|
||||||
|
>= self._context['test_limit']):
|
||||||
|
return
|
||||||
|
|
||||||
|
self._logger.debug(f"Running Job: {variantLabel}")
|
||||||
|
|
||||||
|
|
||||||
|
# Phase 1: Setup source files
|
||||||
|
|
||||||
|
if not variantBasename:
|
||||||
|
raise ValueError(f"{variantLabel}: Testfile basename is falsy")
|
||||||
|
|
||||||
|
|
||||||
|
self.clearTestDirectory()
|
||||||
|
|
||||||
|
self._logger.debug(f"Creating test file: {variantFilename}")
|
||||||
|
mediaFilePath = createMediaTestFile(mediaDescriptor=sourceMediaDescriptor,
|
||||||
|
baseName=variantBasename,
|
||||||
|
directory=self._testDirectory,
|
||||||
|
logger=self._logger,
|
||||||
|
length = 2)
|
||||||
|
|
||||||
|
|
||||||
|
# Phase 2: Run ffx
|
||||||
|
|
||||||
|
commandSequence = [sys.executable,
|
||||||
|
self._ffxExecutablePath]
|
||||||
|
|
||||||
|
if self._context['verbosity']:
|
||||||
|
commandSequence += ['--verbose',
|
||||||
|
str(self._context['verbosity'])]
|
||||||
|
|
||||||
|
commandSequence += ['convert',
|
||||||
|
mediaFilePath,
|
||||||
|
'--no-prompt',
|
||||||
|
'--no-signature']
|
||||||
|
|
||||||
|
if variantFilenameLabel:
|
||||||
|
commandSequence += ['--label', variantFilenameLabel]
|
||||||
|
|
||||||
|
|
||||||
|
commandSequence += ['--no-pattern']
|
||||||
|
commandSequence += ['--no-tmdb']
|
||||||
|
|
||||||
|
out, err, rc = executeProcess(commandSequence, directory = self._testDirectory, context = self._context)
|
||||||
|
|
||||||
|
if out and self._context['verbosity'] >= 9:
|
||||||
|
self._logger.debug(f"{variantLabel}: Process output: {out}")
|
||||||
|
if rc:
|
||||||
|
self._logger.debug(f"{variantLabel}: Process returned ERROR {rc} ({err})")
|
||||||
|
|
||||||
|
|
||||||
|
# Phase 3: Evaluate results
|
||||||
|
|
||||||
|
resultFilenames = [rf for rf in self.getFilenamesInTestDirectory() if rf != 'ffmpeg2pass-0.log' and rf != variantFilename]
|
||||||
|
self._logger.debug(f"{variantLabel}: Result filenames: {resultFilenames}")
|
||||||
|
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
jobFailed = bool(rc)
|
||||||
|
|
||||||
|
self._logger.debug(f"{variantLabel}: Should fail: {shouldFail} / actually failed: {jobFailed}")
|
||||||
|
|
||||||
|
assert (jobFailed == shouldFail
|
||||||
|
), f"Process {'failed' if jobFailed else 'did not fail'}"
|
||||||
|
|
||||||
|
if not jobFailed:
|
||||||
|
|
||||||
|
expectedResultFilePath = os.path.join(self._testDirectory, f"{expectedFilename}")
|
||||||
|
|
||||||
|
assert (os.path.isfile(expectedResultFilePath)
|
||||||
|
), f"Result file {expectedFilename} in path '{self._testDirectory}' wasn't created"
|
||||||
|
|
||||||
|
for assertIndex in range(len(assertSelectorList)):
|
||||||
|
|
||||||
|
assertSelector = assertSelectorList[assertIndex]
|
||||||
|
assertFunc = assertFuncList[assertIndex]
|
||||||
|
assertVariant = variantList[assertIndex]
|
||||||
|
|
||||||
|
if assertSelector == 'B':
|
||||||
|
#TODO: per file find
|
||||||
|
testObj = {'filenames': resultFilenames}
|
||||||
|
assertFunc(testObj=testObj)
|
||||||
|
if assertSelector == 'L':
|
||||||
|
assertFunc()
|
||||||
|
if assertSelector == 'I':
|
||||||
|
assertFunc()
|
||||||
|
|
||||||
|
self._context['test_passed_counter'] += 1
|
||||||
|
self._reportLogger.info(f"{variantLabel}: Test passed")
|
||||||
|
|
||||||
|
except AssertionError as ae:
|
||||||
|
|
||||||
|
self._context['test_failed_counter'] += 1
|
||||||
|
self._reportLogger.error(f"{variantLabel}: Test FAILED ({ae})")
|
||||||
|
|
||||||
|
|
||||||
|
def run(self):
|
||||||
|
for BC in BasenameCombinator.getAllClassReferences():
|
||||||
|
self._logger.debug(f"BC={BC.__name__}")
|
||||||
|
bc = BC(context = self._context)
|
||||||
|
for y in bc.getYield():
|
||||||
|
self.job(y)
|
||||||
165
src/ffx/test/scenario_2.py
Normal file
165
src/ffx/test/scenario_2.py
Normal file
@@ -0,0 +1,165 @@
|
|||||||
|
import os, sys, click
|
||||||
|
|
||||||
|
from .scenario import Scenario
|
||||||
|
|
||||||
|
from ffx.test.helper import createMediaTestFile
|
||||||
|
from ffx.process import executeProcess
|
||||||
|
|
||||||
|
from ffx.file_properties import FileProperties
|
||||||
|
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
|
||||||
|
from ffx.test.media_combinator import MediaCombinator
|
||||||
|
|
||||||
|
|
||||||
|
class Scenario2(Scenario):
|
||||||
|
"""Creating file VAa, h264/aac/aac
|
||||||
|
Converting to VaA, vp9/opus/opus
|
||||||
|
No tmdb, default parameters"""
|
||||||
|
|
||||||
|
TEST_FILE_EXTENSION = 'mkv'
|
||||||
|
EXPECTED_FILE_EXTENSION = 'webm'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context):
|
||||||
|
|
||||||
|
context['use_tmdb'] = False
|
||||||
|
context['use_pattern'] = False
|
||||||
|
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
def getScenario(self):
|
||||||
|
return self.__class__.__name__[8:]
|
||||||
|
|
||||||
|
|
||||||
|
def job(self, yieldObj: dict):
|
||||||
|
|
||||||
|
testContext = self._context.copy()
|
||||||
|
|
||||||
|
targetYieldObj = yieldObj['target']
|
||||||
|
# presetYieldObj = yieldObj['preset'] # not used here
|
||||||
|
|
||||||
|
identifier = targetYieldObj['identifier']
|
||||||
|
variantList = targetYieldObj['variants']
|
||||||
|
|
||||||
|
variantIdentifier = '-'.join(variantList)
|
||||||
|
variantLabel = f"{self.__class__.__name__} Variant {variantIdentifier}"
|
||||||
|
|
||||||
|
sourceMediaDescriptor: MediaDescriptor = targetYieldObj['payload']
|
||||||
|
|
||||||
|
assertSelectorList: list = targetYieldObj['assertSelectors']
|
||||||
|
assertFuncList = targetYieldObj['assertFuncs']
|
||||||
|
shouldFail = targetYieldObj['shouldFail']
|
||||||
|
|
||||||
|
|
||||||
|
if self._context['test_variant'] and not variantIdentifier.startswith(self._context['test_variant']):
|
||||||
|
return
|
||||||
|
|
||||||
|
if (self._context['test_limit'] and (self._context['test_passed_counter'] + self._context['test_failed_counter'])
|
||||||
|
>= self._context['test_limit']):
|
||||||
|
return
|
||||||
|
|
||||||
|
self._logger.debug(f"Running Job: {variantLabel}")
|
||||||
|
|
||||||
|
|
||||||
|
# Phase 1: Setup source files
|
||||||
|
|
||||||
|
self.clearTestDirectory()
|
||||||
|
mediaFilePath = createMediaTestFile(mediaDescriptor=sourceMediaDescriptor,
|
||||||
|
directory=self._testDirectory,
|
||||||
|
logger=self._logger,
|
||||||
|
length = 2)
|
||||||
|
|
||||||
|
|
||||||
|
# Phase 2: Run ffx
|
||||||
|
|
||||||
|
commandSequence = [sys.executable,
|
||||||
|
self._ffxExecutablePath]
|
||||||
|
|
||||||
|
if self._context['verbosity']:
|
||||||
|
commandSequence += ['--verbose',
|
||||||
|
str(self._context['verbosity'])]
|
||||||
|
|
||||||
|
commandSequence += ['convert',
|
||||||
|
mediaFilePath,
|
||||||
|
'--no-prompt',
|
||||||
|
'--no-signature']
|
||||||
|
|
||||||
|
out, err, rc = executeProcess(commandSequence, directory = self._testDirectory, context = self._context)
|
||||||
|
|
||||||
|
if out and self._context['verbosity'] >= 9:
|
||||||
|
self._logger.debug(f"{variantLabel}: Process output: {out}")
|
||||||
|
if rc:
|
||||||
|
self._logger.debug(f"{variantLabel}: Process returned ERROR {rc} ({err})")
|
||||||
|
|
||||||
|
|
||||||
|
# Phase 3: Evaluate results
|
||||||
|
|
||||||
|
resultFilenames = [rf for rf in self.getFilenamesInTestDirectory() if rf.endswith(f".{Scenario2.EXPECTED_FILE_EXTENSION}")]
|
||||||
|
|
||||||
|
self._logger.debug(f"{variantLabel}: Result filenames: {resultFilenames}")
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
jobFailed = bool(rc)
|
||||||
|
self._logger.debug(f"{variantLabel}: Should fail: {shouldFail} / actually failed: {jobFailed}")
|
||||||
|
|
||||||
|
assert (jobFailed == shouldFail
|
||||||
|
), f"Process {'failed' if jobFailed else 'did not fail'}"
|
||||||
|
|
||||||
|
|
||||||
|
if not jobFailed:
|
||||||
|
|
||||||
|
resultFile = os.path.join(self._testDirectory, 'media.webm')
|
||||||
|
|
||||||
|
assert (os.path.isfile(resultFile)
|
||||||
|
), f"Result file 'media.webm' in path '{self._testDirectory}' wasn't created"
|
||||||
|
|
||||||
|
resultFileProperties = FileProperties(testContext, resultFile)
|
||||||
|
resultMediaDescriptor = resultFileProperties.getMediaDescriptor()
|
||||||
|
|
||||||
|
resultMediaTracks = resultMediaDescriptor.getAllTrackDescriptors()
|
||||||
|
|
||||||
|
for assertIndex in range(len(assertSelectorList)):
|
||||||
|
|
||||||
|
assertSelector = assertSelectorList[assertIndex]
|
||||||
|
assertFunc = assertFuncList[assertIndex]
|
||||||
|
assertVariant = variantList[assertIndex]
|
||||||
|
|
||||||
|
if assertSelector == 'M':
|
||||||
|
assertFunc()
|
||||||
|
for variantIndex in range(len(assertVariant)):
|
||||||
|
assert (assertVariant[variantIndex].lower() == resultMediaTracks[variantIndex].getType().indicator()
|
||||||
|
), f"Stream #{variantIndex} is not of type {resultMediaTracks[variantIndex].getType().label()}"
|
||||||
|
|
||||||
|
elif assertSelector == 'AD' or assertSelector == 'AT':
|
||||||
|
assertFunc({'tracks': resultMediaDescriptor.getAudioTracks()})
|
||||||
|
|
||||||
|
elif assertSelector == 'SD' or assertSelector == 'ST':
|
||||||
|
assertFunc({'tracks': resultMediaDescriptor.getSubtitleTracks()})
|
||||||
|
|
||||||
|
elif type(assertSelector) is str:
|
||||||
|
if assertSelector == 'J':
|
||||||
|
assertFunc()
|
||||||
|
|
||||||
|
|
||||||
|
self._context['test_passed_counter'] += 1
|
||||||
|
self._reportLogger.info(f"{variantLabel}: Test passed")
|
||||||
|
|
||||||
|
except AssertionError as ae:
|
||||||
|
|
||||||
|
self._context['test_failed_counter'] += 1
|
||||||
|
self._reportLogger.error(f"{variantLabel}: Test FAILED ({ae})")
|
||||||
|
|
||||||
|
|
||||||
|
def run(self):
|
||||||
|
MC_list = MediaCombinator.getAllClassReferences()
|
||||||
|
for MC in MC_list:
|
||||||
|
self._logger.debug(f"MC={MC.__name__}")
|
||||||
|
mc = MC(context = self._context)
|
||||||
|
for y in mc.getYield():
|
||||||
|
self.job(y)
|
||||||
274
src/ffx/test/scenario_4.py
Normal file
274
src/ffx/test/scenario_4.py
Normal file
@@ -0,0 +1,274 @@
|
|||||||
|
import os, sys, click
|
||||||
|
|
||||||
|
from .scenario import Scenario
|
||||||
|
|
||||||
|
from ffx.test.helper import createMediaTestFile
|
||||||
|
from ffx.process import executeProcess
|
||||||
|
from ffx.database import databaseContext
|
||||||
|
|
||||||
|
from ffx.test.helper import createEmptyDirectory
|
||||||
|
from ffx.helper import getEpisodeFileBasename
|
||||||
|
|
||||||
|
from ffx.file_properties import FileProperties
|
||||||
|
|
||||||
|
from ffx.media_descriptor import MediaDescriptor
|
||||||
|
from ffx.track_descriptor import TrackDescriptor
|
||||||
|
|
||||||
|
from ffx.track_type import TrackType
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
|
||||||
|
from ffx.test.media_combinator import MediaCombinator
|
||||||
|
from ffx.test.indicator_combinator import IndicatorCombinator
|
||||||
|
|
||||||
|
from ffx.show_descriptor import ShowDescriptor
|
||||||
|
|
||||||
|
|
||||||
|
from ffx.tmdb_controller import TmdbController
|
||||||
|
from ffx.tmdb_controller import TMDB_API_KEY_NOT_PRESENT_EXCEPTION
|
||||||
|
|
||||||
|
class Scenario4(Scenario):
|
||||||
|
|
||||||
|
TEST_SHOW_IDENTIFIER = 83095
|
||||||
|
TEST_SHOW_NAME = 'The Rising of the Shield Hero'
|
||||||
|
TEST_SHOW_YEAR = 2019
|
||||||
|
|
||||||
|
TEST_FILE_LABEL = 'rotsh'
|
||||||
|
TEST_FILE_EXTENSION = 'mkv'
|
||||||
|
|
||||||
|
TEST_PATTERN = f"{TEST_FILE_LABEL}_{FileProperties.SE_INDICATOR_PATTERN}.{TEST_FILE_EXTENSION}"
|
||||||
|
|
||||||
|
EXPECTED_FILE_EXTENSION = 'webm'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__tmdbApiKey = os.environ.get('TMDB_API_KEY', None)
|
||||||
|
if self.__tmdbApiKey is None:
|
||||||
|
raise TMDB_API_KEY_NOT_PRESENT_EXCEPTION
|
||||||
|
|
||||||
|
self.__ic = IndicatorCombinator(context = context)
|
||||||
|
self.__tc = TmdbController()
|
||||||
|
|
||||||
|
|
||||||
|
kwargs = {}
|
||||||
|
kwargs[ShowDescriptor.ID_KEY] = Scenario4.TEST_SHOW_IDENTIFIER
|
||||||
|
kwargs[ShowDescriptor.NAME_KEY] = Scenario4.TEST_SHOW_NAME
|
||||||
|
kwargs[ShowDescriptor.YEAR_KEY] = Scenario4.TEST_SHOW_YEAR
|
||||||
|
|
||||||
|
self.__testShowDescriptor = ShowDescriptor(**kwargs)
|
||||||
|
|
||||||
|
def getScenario(self):
|
||||||
|
return self.__class__.__name__[8:]
|
||||||
|
|
||||||
|
|
||||||
|
def prepareTestDatabase(self, sourceMediaDescriptor: MediaDescriptor):
|
||||||
|
|
||||||
|
if not self._context['database'] is None:
|
||||||
|
self._context['database']['engine'].dispose()
|
||||||
|
|
||||||
|
if os.path.isfile(self._testDbFilePath):
|
||||||
|
os.unlink(self._testDbFilePath)
|
||||||
|
self._context['database'] = None
|
||||||
|
|
||||||
|
self._logger.debug(f"Creating test db with path {self._testDbFilePath}")
|
||||||
|
self._context['database'] = databaseContext(databasePath=self._testDbFilePath)
|
||||||
|
|
||||||
|
|
||||||
|
self._logger.debug(f"Adding test show '{self.__testShowDescriptor.getFilenamePrefix()}' to test db")
|
||||||
|
if not self._sc.updateShow(self.__testShowDescriptor):
|
||||||
|
raise click.ClickException('Could not create test show in db')
|
||||||
|
|
||||||
|
testPatternDescriptor = {
|
||||||
|
'show_id': Scenario4.TEST_SHOW_IDENTIFIER,
|
||||||
|
'pattern': Scenario4.TEST_PATTERN
|
||||||
|
}
|
||||||
|
patternId = self._pc.addPattern(testPatternDescriptor)
|
||||||
|
|
||||||
|
if patternId:
|
||||||
|
self._mc.setPatternMediaDescriptor(sourceMediaDescriptor, patternId)
|
||||||
|
|
||||||
|
|
||||||
|
def job(self, yieldObj: dict):
|
||||||
|
|
||||||
|
testContext = self._context.copy()
|
||||||
|
|
||||||
|
if 'preset' not in yieldObj.keys():
|
||||||
|
raise KeyError('yieldObj did not contain presets')
|
||||||
|
|
||||||
|
targetYieldObj = yieldObj['target']
|
||||||
|
presetYieldObj = yieldObj['preset']
|
||||||
|
|
||||||
|
identifier = targetYieldObj['identifier']
|
||||||
|
variantList = targetYieldObj['variants']
|
||||||
|
|
||||||
|
variantIdentifier = '-'.join(variantList)
|
||||||
|
variantLabel = f"{self.__class__.__name__} Variant {variantIdentifier}"
|
||||||
|
|
||||||
|
sourceMediaDescriptor: MediaDescriptor = targetYieldObj['payload']
|
||||||
|
presetMediaDescriptor: MediaDescriptor = presetYieldObj['payload']
|
||||||
|
|
||||||
|
assertSelectorList: list = presetYieldObj['assertSelectors']
|
||||||
|
assertFuncList = presetYieldObj['assertFuncs']
|
||||||
|
shouldFail = presetYieldObj['shouldFail']
|
||||||
|
|
||||||
|
|
||||||
|
if self._context['test_variant'] and not variantIdentifier.startswith(self._context['test_variant']):
|
||||||
|
return
|
||||||
|
|
||||||
|
if (self._context['test_limit'] and (self._context['test_passed_counter'] + self._context['test_failed_counter'])
|
||||||
|
>= self._context['test_limit']):
|
||||||
|
return
|
||||||
|
|
||||||
|
self._logger.debug(f"Running Job: {variantLabel}")
|
||||||
|
|
||||||
|
|
||||||
|
for l in presetMediaDescriptor.getConfiguration(label = 'presetMediaDescriptor'):
|
||||||
|
self._logger.debug(l)
|
||||||
|
|
||||||
|
for l in sourceMediaDescriptor.getConfiguration(label = 'sourceMediaDescriptor'):
|
||||||
|
self._logger.debug(l)
|
||||||
|
|
||||||
|
|
||||||
|
# Phase 1: Setup source files
|
||||||
|
|
||||||
|
self.clearTestDirectory()
|
||||||
|
|
||||||
|
testFileList = []
|
||||||
|
for indicatorObj in [y for y in self.__ic.getYield() if y['indicator']]:
|
||||||
|
|
||||||
|
indicator = indicatorObj['indicator']
|
||||||
|
|
||||||
|
testFileObj = {}
|
||||||
|
testFileObj['season'] = indicatorObj['season']
|
||||||
|
testFileObj['episode'] = indicatorObj['episode']
|
||||||
|
|
||||||
|
testFileObj['basename'] = f"{Scenario4.TEST_FILE_LABEL}_{indicator}"
|
||||||
|
|
||||||
|
testFileObj['path'] = createMediaTestFile(mediaDescriptor = presetMediaDescriptor,
|
||||||
|
directory = self._testDirectory,
|
||||||
|
baseName = testFileObj['basename'],
|
||||||
|
logger=self._logger,
|
||||||
|
length = 2)
|
||||||
|
testFileObj['filename'] = f"{testFileObj['basename']}.{Scenario4.TEST_FILE_EXTENSION}"
|
||||||
|
|
||||||
|
testFileList.append(testFileObj)
|
||||||
|
|
||||||
|
|
||||||
|
# Phase 2: Prepare database
|
||||||
|
|
||||||
|
self.createEmptyTestDatabase()
|
||||||
|
self.prepareTestDatabase(sourceMediaDescriptor)
|
||||||
|
|
||||||
|
|
||||||
|
# Phase 3: Run ffx
|
||||||
|
|
||||||
|
commandSequence = [sys.executable,
|
||||||
|
self._ffxExecutablePath]
|
||||||
|
|
||||||
|
if self._context['verbosity']:
|
||||||
|
commandSequence += ['--verbose',
|
||||||
|
str(self._context['verbosity'])]
|
||||||
|
|
||||||
|
commandSequence += ['--database-file',
|
||||||
|
self._testDbFilePath,
|
||||||
|
'convert']
|
||||||
|
commandSequence += [tfo['filename'] for tfo in testFileList]
|
||||||
|
|
||||||
|
commandSequence += ['--no-prompt', '--no-signature']
|
||||||
|
|
||||||
|
out, err, rc = executeProcess(commandSequence, directory = self._testDirectory, context = self._context)
|
||||||
|
|
||||||
|
if out and self._context['verbosity'] >= 9:
|
||||||
|
self._logger.debug(f"{variantLabel}: Process output: {out}")
|
||||||
|
if rc:
|
||||||
|
self._logger.debug(f"{variantLabel}: Process returned ERROR {rc} ({err})")
|
||||||
|
|
||||||
|
|
||||||
|
# Phase 4: Evaluate results
|
||||||
|
|
||||||
|
resultFilenames = [rf for rf in self.getFilenamesInTestDirectory() if rf.endswith(f".{Scenario4.EXPECTED_FILE_EXTENSION}")]
|
||||||
|
|
||||||
|
self._logger.debug(f"{variantLabel}: Result filenames: {resultFilenames}")
|
||||||
|
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
jobFailed = bool(rc)
|
||||||
|
self._logger.debug(f"{variantLabel}: Should fail: {shouldFail} / actually failed: {jobFailed}")
|
||||||
|
|
||||||
|
assert (jobFailed == shouldFail
|
||||||
|
), f"Process {'failed' if jobFailed else 'did not fail'}"
|
||||||
|
|
||||||
|
if not jobFailed:
|
||||||
|
|
||||||
|
for tfo in testFileList:
|
||||||
|
|
||||||
|
tmdbEpisodeResult = self.__tc.queryEpisode(Scenario4.TEST_SHOW_IDENTIFIER,
|
||||||
|
tfo['season'], tfo['episode'])
|
||||||
|
|
||||||
|
expectedFileBasename = getEpisodeFileBasename(self.__testShowDescriptor.getFilenamePrefix(),
|
||||||
|
tmdbEpisodeResult['name'],
|
||||||
|
tfo['season'], tfo['episode'],
|
||||||
|
context=testContext)
|
||||||
|
|
||||||
|
expectedFilename = f"{expectedFileBasename}.{Scenario4.EXPECTED_FILE_EXTENSION}"
|
||||||
|
expectedFilePath = os.path.join(self._testDirectory, expectedFilename)
|
||||||
|
|
||||||
|
assert (os.path.isfile(expectedFilePath)
|
||||||
|
), f"Result file '{expectedFilename}' in path '{self._testDirectory}' wasn't created"
|
||||||
|
|
||||||
|
|
||||||
|
rfp = FileProperties(testContext, expectedFilePath)
|
||||||
|
self._logger.debug(f"{variantLabel}: Result file properties: {rfp.getFilename()} season={rfp.getSeason()} episode={rfp.getEpisode()}")
|
||||||
|
|
||||||
|
rmd = rfp.getMediaDescriptor()
|
||||||
|
rmt = rmd.getAllTrackDescriptors()
|
||||||
|
|
||||||
|
for l in rmd.getConfiguration(label = 'resultMediaDescriptor'):
|
||||||
|
self._logger.debug(l)
|
||||||
|
|
||||||
|
# num tracks differ
|
||||||
|
rmd.applySourceIndices(sourceMediaDescriptor)
|
||||||
|
|
||||||
|
|
||||||
|
for assertIndex in range(len(assertSelectorList)):
|
||||||
|
|
||||||
|
assertSelector = assertSelectorList[assertIndex]
|
||||||
|
assertFunc = assertFuncList[assertIndex]
|
||||||
|
assertVariant = variantList[assertIndex]
|
||||||
|
|
||||||
|
if assertSelector == 'M':
|
||||||
|
assertFunc()
|
||||||
|
for variantIndex in range(len(assertVariant)):
|
||||||
|
assert (assertVariant[variantIndex].lower() == rmt[variantIndex].getType().indicator()
|
||||||
|
), f"Stream #{variantIndex} is not of type {rmt[variantIndex].getType().label()}"
|
||||||
|
|
||||||
|
if assertSelector == 'AD' or assertSelector == 'AT':
|
||||||
|
assertFunc({'tracks': rmd.getAudioTracks()})
|
||||||
|
|
||||||
|
elif assertSelector == 'SD' or assertSelector == 'ST':
|
||||||
|
assertFunc({'tracks': rmd.getSubtitleTracks()})
|
||||||
|
|
||||||
|
elif type(assertSelector) is str:
|
||||||
|
if assertSelector == 'J':
|
||||||
|
assertFunc()
|
||||||
|
|
||||||
|
|
||||||
|
self._context['test_passed_counter'] += 1
|
||||||
|
self._reportLogger.info(f"\n{variantLabel}: Test passed\n")
|
||||||
|
|
||||||
|
except AssertionError as ae:
|
||||||
|
|
||||||
|
self._context['test_failed_counter'] += 1
|
||||||
|
self._reportLogger.error(f"\n{variantLabel}: Test FAILED ({ae})\n")
|
||||||
|
|
||||||
|
|
||||||
|
def run(self):
|
||||||
|
|
||||||
|
MC_list = [MediaCombinator.getClassReference(6)]
|
||||||
|
for MC in MC_list:
|
||||||
|
self._logger.debug(f"MC={MC.__name__}")
|
||||||
|
mc = MC(context = self._context, createPresets = True)
|
||||||
|
for y in mc.getYield():
|
||||||
|
self.job(y)
|
||||||
33
src/ffx/test/show_combinator.py
Normal file
33
src/ffx/test/show_combinator.py
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
class ShowCombinator():
|
||||||
|
|
||||||
|
IDENTIFIER = 'show'
|
||||||
|
|
||||||
|
SHOW_LIST = [
|
||||||
|
'Boruto; Naruto Next Generations (2017)',
|
||||||
|
'The Rising of the Shield Hero (2019)',
|
||||||
|
'Scrubs - Die Anfänger (2001)'
|
||||||
|
]
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getIdentifier(self):
|
||||||
|
return ShowCombinator.IDENTIFIER
|
||||||
|
|
||||||
|
def getPayload(self, showIndex):
|
||||||
|
return {
|
||||||
|
'variant': f"S{showIndex}",
|
||||||
|
'show': ShowCombinator.SHOW_LIST[showIndex]
|
||||||
|
}
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
for showIndex in range(len(ShowCombinator.SHOW_LIST)):
|
||||||
|
yield self.getPayload(showIndex)
|
||||||
33
src/ffx/test/title_combinator.py
Normal file
33
src/ffx/test/title_combinator.py
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
class TitleCombinator():
|
||||||
|
|
||||||
|
IDENTIFIER = 'title'
|
||||||
|
|
||||||
|
TITLE_LIST = [
|
||||||
|
'The Sound of Space',
|
||||||
|
'2001; Odyssee im Weltraum (1968)',
|
||||||
|
'Ansible 101'
|
||||||
|
]
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getIdentifier(self):
|
||||||
|
return TitleCombinator.IDENTIFIER
|
||||||
|
|
||||||
|
def getPayload(self, titleIndex):
|
||||||
|
return {
|
||||||
|
'variant': f"S{titleIndex}",
|
||||||
|
'show': TitleCombinator.TITLE_LIST[titleIndex]
|
||||||
|
}
|
||||||
|
|
||||||
|
def assertFunc(self, testObj = {}):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
def getYield(self):
|
||||||
|
for titleIndex in range(len(TitleCombinator.TITLE_LIST)):
|
||||||
|
yield self.getPayload(titleIndex)
|
||||||
33
src/ffx/test/track_tag_combinator_2.py
Normal file
33
src/ffx/test/track_tag_combinator_2.py
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
import os, sys, importlib, glob, inspect, itertools
|
||||||
|
|
||||||
|
class TrackTagCombinator2():
|
||||||
|
|
||||||
|
IDENTIFIER = 'trackTag2'
|
||||||
|
|
||||||
|
def __init__(self, context = None):
|
||||||
|
self._context = context
|
||||||
|
self._logger = context['logger']
|
||||||
|
self._reportLogger = context['report_logger']
|
||||||
|
|
||||||
|
def getIdentifier(self):
|
||||||
|
return TrackTagCombinator2.IDENTIFIER
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def list():
|
||||||
|
basePath = os.path.dirname(__file__)
|
||||||
|
return [os.path.basename(p)[23:-3]
|
||||||
|
for p
|
||||||
|
in glob.glob(f"{ basePath }/track_tag_combinator_2_*.py", recursive = True)
|
||||||
|
if p != __file__]
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getClassReference(identifier):
|
||||||
|
importlib.import_module(f"ffx.test.track_tag_combinator_2_{ identifier }")
|
||||||
|
for name, obj in inspect.getmembers(sys.modules[f"ffx.test.track_tag_combinator_2_{ identifier }"]):
|
||||||
|
#HINT: Excluding DispositionCombination as it seems to be included by import (?)
|
||||||
|
if inspect.isclass(obj) and name != 'TrackTagCombinator2' and name.startswith('TrackTagCombinator2'):
|
||||||
|
return obj
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def getAllClassReferences():
|
||||||
|
return [TrackTagCombinator2.getClassReference(i) for i in TrackTagCombinator2.list()]
|
||||||
89
src/ffx/test/track_tag_combinator_2_0.py
Normal file
89
src/ffx/test/track_tag_combinator_2_0.py
Normal file
@@ -0,0 +1,89 @@
|
|||||||
|
import os, sys, importlib, glob, inspect
|
||||||
|
|
||||||
|
from ffx.track_disposition import TrackDisposition
|
||||||
|
from .track_tag_combinator_2 import TrackTagCombinator2
|
||||||
|
|
||||||
|
|
||||||
|
class TrackTagCombinator20(TrackTagCombinator2):
|
||||||
|
|
||||||
|
VARIANT = 'T00'
|
||||||
|
|
||||||
|
|
||||||
|
def __init__(self, context = None,
|
||||||
|
createPresets: bool = False):
|
||||||
|
super().__init__(context)
|
||||||
|
|
||||||
|
self.__createPresets = createPresets
|
||||||
|
|
||||||
|
def getVariant(self):
|
||||||
|
return TrackTagCombinator20.VARIANT
|
||||||
|
|
||||||
|
|
||||||
|
def getPayload(self):
|
||||||
|
|
||||||
|
subTrack0 = {}
|
||||||
|
subTrack1 = {}
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
subTrack0['THIS_IS'] = 'track0'
|
||||||
|
subTrack1['THIS_IS'] = 'track1'
|
||||||
|
|
||||||
|
return (subTrack0,
|
||||||
|
subTrack1)
|
||||||
|
|
||||||
|
|
||||||
|
def createAssertFunc(self):
|
||||||
|
|
||||||
|
if self.__createPresets:
|
||||||
|
|
||||||
|
def f(assertObj: dict = {}):
|
||||||
|
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
resortedTrackDescriptors = sorted(assertObj['tracks'], key=lambda d: d.getSourceIndex())
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (not 'language' in resortedTrackDescriptors[0].getTags().keys()
|
||||||
|
), f"Stream src_index={resortedTrackDescriptors[0].getSourceIndex()} index={resortedTrackDescriptors[0].getIndex()} [{resortedTrackDescriptors[0].getType().label()}:{resortedTrackDescriptors[0].getSubIndex()}] has set language tag"
|
||||||
|
assert (not 'title' in resortedTrackDescriptors[0].getTags().keys()
|
||||||
|
), f"Stream src_index={resortedTrackDescriptors[0].getSourceIndex()} index={resortedTrackDescriptors[0].getIndex()} [{resortedTrackDescriptors[0].getType().label()}:{resortedTrackDescriptors[0].getSubIndex()}] has set title tag"
|
||||||
|
|
||||||
|
assert ('THIS_IS' in resortedTrackDescriptors[0].getTags().keys() and resortedTrackDescriptors[0].getTags()['THIS_IS'] == 'track0'
|
||||||
|
), f"Stream src_index={resortedTrackDescriptors[0].getSourceIndex()} index={resortedTrackDescriptors[0].getIndex()} [{resortedTrackDescriptors[0].getType().label()}:{resortedTrackDescriptors[0].getSubIndex()}] has not preserved tag THIS_IS"
|
||||||
|
|
||||||
|
# source subIndex 1
|
||||||
|
assert (not 'language' in resortedTrackDescriptors[1].getTags().keys()
|
||||||
|
), f"Stream src_index={resortedTrackDescriptors[1].getSourceIndex()} index={resortedTrackDescriptors[1].getIndex()} [{resortedTrackDescriptors[1].getType().label()}:{resortedTrackDescriptors[1].getSubIndex()}] has set language tag"
|
||||||
|
assert (not 'title' in resortedTrackDescriptors[1].getTags().keys()
|
||||||
|
), f"Stream src_index={resortedTrackDescriptors[1].getSourceIndex()} index={resortedTrackDescriptors[1].getIndex()} [{resortedTrackDescriptors[1].getType().label()}:{resortedTrackDescriptors[1].getSubIndex()}] has set title tag"
|
||||||
|
assert (not 'title' in resortedTrackDescriptors[1].getTags().keys()
|
||||||
|
), f"Stream src_index={resortedTrackDescriptors[1].getSourceIndex()} index={resortedTrackDescriptors[1].getIndex()} [{resortedTrackDescriptors[1].getType().label()}:{resortedTrackDescriptors[1].getSubIndex()}] has set title tag"
|
||||||
|
|
||||||
|
assert ('THIS_IS' in resortedTrackDescriptors[1].getTags().keys() and resortedTrackDescriptors[1].getTags()['THIS_IS'] == 'track1'
|
||||||
|
), f"Stream src_index={resortedTrackDescriptors[1].getSourceIndex()} index={resortedTrackDescriptors[1].getIndex()} [{resortedTrackDescriptors[1].getType().label()}:{resortedTrackDescriptors[1].getSubIndex()}] has not preserved tag THIS_IS"
|
||||||
|
|
||||||
|
else:
|
||||||
|
|
||||||
|
def f(assertObj: dict = {}):
|
||||||
|
|
||||||
|
if not 'tracks' in assertObj.keys():
|
||||||
|
raise KeyError("assertObj does not contain key 'tracks'")
|
||||||
|
resortedTrackDescriptors = sorted(assertObj['tracks'], key=lambda d: d.getSourceIndex())
|
||||||
|
|
||||||
|
# source subIndex 0
|
||||||
|
assert (not 'language' in resortedTrackDescriptors[0].getTags().keys()
|
||||||
|
), f"Stream src_index={resortedTrackDescriptors[0].getSourceIndex()} index={resortedTrackDescriptors[0].getIndex()} [{resortedTrackDescriptors[0].getType().label()}:{resortedTrackDescriptors[0].getSubIndex()}] has set language tag"
|
||||||
|
assert (not 'title' in resortedTrackDescriptors[0].getTags().keys()
|
||||||
|
), f"Stream src_index={resortedTrackDescriptors[0].getSourceIndex()} index={resortedTrackDescriptors[0].getIndex()} [{resortedTrackDescriptors[0].getType().label()}:{resortedTrackDescriptors[0].getSubIndex()}] has set title tag"
|
||||||
|
|
||||||
|
# source subIndex 1
|
||||||
|
assert (not 'language' in resortedTrackDescriptors[1].getTags().keys()
|
||||||
|
), f"Stream src_index={resortedTrackDescriptors[1].getSourceIndex()} index={resortedTrackDescriptors[1].getIndex()} [{resortedTrackDescriptors[1].getType().label()}:{resortedTrackDescriptors[1].getSubIndex()}] has set language tag"
|
||||||
|
assert (not 'title' in resortedTrackDescriptors[1].getTags().keys()
|
||||||
|
), f"Stream src_index={resortedTrackDescriptors[1].getSourceIndex()} index={resortedTrackDescriptors[1].getIndex()} [{resortedTrackDescriptors[1].getType().label()}:{resortedTrackDescriptors[1].getSubIndex()}] has set title tag"
|
||||||
|
|
||||||
|
return f
|
||||||
|
|
||||||
|
|
||||||
|
def shouldFail(self):
|
||||||
|
return False
|
||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user