Compare commits

..

No commits in common. "40eb0f3518b47b1027d5ff5c4e2cad0e241e8d65" and "a479cc8c392bae9c9a82994258e872f64992d401" have entirely different histories.

141
main.py
View File

@ -3,124 +3,49 @@ from bs4 import BeautifulSoup
import requests
from simple_term_menu import TerminalMenu
import os
import sys
import subprocess
from operator import itemgetter
def get_files(url: str) -> []:
if (url in ["/../", "../"]):
return ([])
response = requests.get(url)
# print("connection:", response.reason)
if (response.status_code != 200):
exit
soup = BeautifulSoup(response.text, 'html.parser')
files = []
for element in soup.findAll("a"):
file = {}
file.update({"name": element.text})
file.update({"link": element["href"]})
files.append(file)
return (files)
db = TinyDB("./database.json", indent=4)
query = Query()
def get_uri(url: str) -> []:
if (url in ["/../", "../"]):
return ([])
response = requests.get(url)
# print("connection:", response.reason)
if (response.status_code != 200):
exit
soup = BeautifulSoup(response.text, 'html.parser')
return(soup.find("h1").text[9:])
def get(files: [], key: str):
names = []
for file in files:
names.append(file.get(key))
return (names)
def open_vlc(url: str) -> None:
with open(os.devnull, 'wb') as devnull:
subprocess.check_call(['vlc', url], stdout=devnull, stderr=subprocess.STDOUT)
def preview(filename: str) -> str:
if (not filename.endswith("/")):
return (None)
files = get_files(url + filename)
return ("\n".join(get(files, "name")))
def files_navigator():
global url
path = cookies.get("path")
if (path):
url + path
else:
path = "/"
while True:
cookies.update({"last_path": path})
db.update({"last_path", path})
files = get_files(url)
terminal_menu = TerminalMenu(get(files, "name"), preview_command=preview, preview_size=0.3, show_search_hint=True, title=get_uri(url))
file_choose = terminal_menu.show()
if (file_choose == None):
sys.exit(0)
elif (file_choose == 0 and get_uri(url) == "/"):
return
file = get(files, "link")[file_choose]
if (not file.endswith("/")):
open_vlc(url + file)
else:
url = url + file
def add_site():
sites = db.table("sites")
nb_sites = len(sites.all());
site_choose = 0;
if (nb_sites == 0):
print("add a site:")
site = {}
site.update({"url": input("url:")})
site.update({"user": input("user(leave blank):")})
site.update({"password": input("password(leave blank):")})
site.update({"id": len(sites_table.all)})
name = input("name[url] :")
if (name == ""):
name = site.get("url")
site.update({"name": name})
sites_table.insert(site)
def set_url(site):
global url
cookies.update({"last_site": site})
if (site.get("user") and site.get("password")):
url = f"http://{site.get('user')}:{site.get('password')}@{site.get('url')}/"
sites.insert(site)
elif (nb_sites != 1):
lst = []
for site in sites.all():
append(sites.get("name"))
terminal_menu = TerminalMenu(lst)
site_choose = terminal_menu.show()
site = sites.all()[site_choose]
if (site.get("user") and site.get("password")):
url = f"http://{site.get('user')}:{site.get('password')}@{site.get('url')}/"
else:
url = f"http://{site.get('url')}/";
while True:
response = requests.get(url)
print("connection:", response.reason)
if (response.status_code != 200):
exit
soup = BeautifulSoup(response.text, 'html.parser')
lst = []
for element in soup.findAll("a"):
lst.append(element.text)
terminal_menu = TerminalMenu(lst)
file = soup.findAll("a")[terminal_menu.show()]["href"]
if (not file.endswith("/")):
with open(os.devnull, 'wb') as devnull:
subprocess.check_call(['vlc', url + file], stdout=devnull, stderr=subprocess.STDOUT)
else:
url = f"http://{site.get('url')}/";
def sites_navigator():
if (len(sites_table.all()) == 0):
add_site()
site = cookies.get("last_site")
if (site):
set_url(site)
files_navigator()
while True:
terminal_menu = TerminalMenu(get(sites_table.all(), "name") + ["", "add", "edit", "remove"], skip_empty_entries=True, show_search_hint=True)
choose = terminal_menu.show()
if (choose == len(sites_table.all()) + 1):
add_site();
elif (choose == len(sites_table.all()) + 2):
pass
elif (choose == len(sites_table.all()) + 3):
pass
else:
set_url(sites_table.all()[choose])
db.update({"last_path", get_uri(url)})
files_navigator()
db = TinyDB("./database.json", indent=4)
query = Query()
sites_table = db.table("sites")
cookies_table = db.table("cookies")
if (len(cookies_table.all()) == 0):
cookies_table.insert({"id": 0})
cookies = cookies_table.all()[0]
sites_navigator();
url = url + file