完成作业一:260326_2509165004

This commit is contained in:
2509165004
2026-03-26 15:59:59 +08:00
parent 9ca58a9ac8
commit 37318d5257
4 changed files with 24 additions and 42 deletions

24
060326-2509165004.py Normal file
View File

@@ -0,0 +1,24 @@
import requests
from bs4 import BeautifulSoup
import time
headers = {'User-Agent':'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/96.0.4664.110 Safari/537.36'}
for i in range(5):
url = f'https://picsum.photos/'
response = requests.get(url, headers=headers,timeout=10)
response.encoding = 'utf-8'
soup = BeautifulSoup(response.text, 'html.parser')
img_tag = soup.select_one('img.resize')
# print(img_tag)
if img_tag:
img_src = img_tag.get('src')
print(f"{i+1}张图片地址:{img_src}")
img_response = requests.get(img_src, timeout=10)
with open('image_{i+1}.jpg', 'wb') as f:
f.write(img_response.content)
print(f"{i+1}张图片保存成功!")

View File

@@ -1,21 +0,0 @@
import requests
from bs4 import BeautifulSoup
headers = {
'User-Agent':'Mozilla/5.0(Windows NT 10.0;Win64;x64)AppleWebKit/537.36(KHTML,like Gecko)Chrome/120.0.0.0 Safari/537.36',
}
url = 'https://movie.douban.com/top250?start=0&filter='
response = requests.get(url,headers=headers,timeout=10)
response.encoding = 'utf-8'
soup = BeautifulSoup(response.text,'html.paser')
movies = []
for a in soup.find_all('a'):
herf = a.get('href','')
if '/subject' in href:
title = a.get_text(strip=True)
print(title)
movies.append(title)
print('------------')
print(movies)

8
lxj.py
View File

@@ -1,8 +0,0 @@
import requests
url = 'http://example.com'
params = {'key':'value'}
response = requests.get(url, params=params)
if response.status_code == 200:
print("请求成功获取到HTML内容")
else:
print(f"请求失败,状态码:{response.status_code}")

View File

@@ -1,13 +0,0 @@
from bs4 import BeautifulSoup
import requests
response = requests.get('https://example.com')
html_content = response.text
soup = BeautifulSoup(html_content,'lxml')
title = soup.find('title').string
print("页面标题:",title)
links = soup.find_all('a')
for link in links:
print("链接地址:",link.get('href'))
div_elements = soup.select('div.class_name')
for div in div_elements:
print("div内容",div.text)