python快速找出大文件

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
import os
import time
from os.path import join, getsize
from heapq import nlargest

def walk_files_and_sizes(start_at: str):
for root, _, files in os.walk(start_at):
for file in files:
path = join(root, file)
try:
size = getsize(path) # bytes
yield path, size
except OSError:
continue

def largest_files(n: int, start_at: str) -> None:
MB = 1024 * 1024
largest = nlargest(n, walk_files_and_sizes(start_at), key=lambda x: x[1])

for path, size in largest:
print(f'{size//MB} MB {path}')

if __name__ == '__main__':
start = time.perf_counter()
largest_files(10, "D:/datasets/")
elapsed = time.perf_counter() - start
print(f'{elapsed} seconds elapsed')
0%