Create iceriktoplama.py
Browse files- iceriktoplama.py +78 -0
iceriktoplama.py
ADDED
|
@@ -0,0 +1,78 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import httpx
|
| 2 |
+
from bs4 import BeautifulSoup
|
| 3 |
+
import json
|
| 4 |
+
from concurrent.futures import ThreadPoolExecutor, as_completed
|
| 5 |
+
from tqdm import tqdm
|
| 6 |
+
import multiprocessing
|
| 7 |
+
import time
|
| 8 |
+
|
| 9 |
+
# URL'leri içeren txt dosyasının yolunu belirtin
|
| 10 |
+
file_path = '/content/url.txt'
|
| 11 |
+
|
| 12 |
+
# URL'leri txt dosyasından okuyun
|
| 13 |
+
with open(file_path, 'r') as file:
|
| 14 |
+
url_list = [line.strip() for line in file if line.strip()]
|
| 15 |
+
|
| 16 |
+
disclaimer = "'zynp_msgdata' veri seti 'sekerlipencere' tarafından hazırlanmıştır."
|
| 17 |
+
cpu_count = multiprocessing.cpu_count()
|
| 18 |
+
max_workers = cpu_count * 300 # İşlemci çekirdek sayısının 20 katı kadar iş parçacığı oluştur
|
| 19 |
+
|
| 20 |
+
def fetch_content(url):
|
| 21 |
+
try:
|
| 22 |
+
with httpx.Client(timeout=5) as client:
|
| 23 |
+
response = client.get(url)
|
| 24 |
+
response.raise_for_status()
|
| 25 |
+
return url, response.text
|
| 26 |
+
except Exception as e:
|
| 27 |
+
print(f"URL alınamadı: {url} - Hata: {e}")
|
| 28 |
+
return url, None
|
| 29 |
+
|
| 30 |
+
def process_content(url_content):
|
| 31 |
+
url, webpage = url_content
|
| 32 |
+
if webpage is None:
|
| 33 |
+
return None
|
| 34 |
+
|
| 35 |
+
try:
|
| 36 |
+
soup = BeautifulSoup(webpage, 'html.parser')
|
| 37 |
+
soru_div = soup.find('div', class_='p-title')
|
| 38 |
+
if soru_div:
|
| 39 |
+
soru = soru_div.get_text(strip=True)
|
| 40 |
+
ayrintili_soru_div = soup.find('div', class_='bbWrapper')
|
| 41 |
+
ayrintili_soru = ayrintili_soru_div.get_text(strip=True) if ayrintili_soru_div else ''
|
| 42 |
+
cevap_divs = soup.find_all('div', class_='bbWrapper')[1:]
|
| 43 |
+
cevaplar = [cevap.get_text(strip=True) for cevap in cevap_divs]
|
| 44 |
+
return {
|
| 45 |
+
'soru': soru,
|
| 46 |
+
'url': url,
|
| 47 |
+
'ayrintili_soru': ayrintili_soru,
|
| 48 |
+
'cevaplar': cevaplar,
|
| 49 |
+
'atıf': disclaimer
|
| 50 |
+
}
|
| 51 |
+
except Exception as e:
|
| 52 |
+
print(f"İçerik işlenirken hata oluştu: {url} - Hata: {e}")
|
| 53 |
+
return None
|
| 54 |
+
|
| 55 |
+
def write_to_json(result):
|
| 56 |
+
if result:
|
| 57 |
+
with open('sonuc.json', 'a', encoding='utf-8') as f:
|
| 58 |
+
json.dump(result, f, ensure_ascii=False, indent=4)
|
| 59 |
+
f.write('\n')
|
| 60 |
+
|
| 61 |
+
def main():
|
| 62 |
+
batch_size = 500 # Toplu işleme için isteklerin gruplanacağı boyut
|
| 63 |
+
with ThreadPoolExecutor(max_workers=max_workers) as executor:
|
| 64 |
+
for i in range(0, len(url_list), batch_size):
|
| 65 |
+
batch_urls = url_list[i:i+batch_size]
|
| 66 |
+
futures = [executor.submit(fetch_content, url) for url in batch_urls]
|
| 67 |
+
with tqdm(total=len(futures), desc="Toplam İşleniyor", unit="URL") as pbar:
|
| 68 |
+
for future in as_completed(futures):
|
| 69 |
+
url_content = future.result()
|
| 70 |
+
result = process_content(url_content)
|
| 71 |
+
write_to_json(result)
|
| 72 |
+
pbar.update(1)
|
| 73 |
+
|
| 74 |
+
if __name__ == "__main__":
|
| 75 |
+
start_time = time.time()
|
| 76 |
+
main()
|
| 77 |
+
end_time = time.time()
|
| 78 |
+
print(f"Toplam süre: {end_time - start_time:.2f} saniye")
|