#!/usr/bin/env python3.4
#coding:utf-8


# from bs4 import BeautifulSoup 
# import  urllib
# import  re

# url= 'http://www.dytt8.net/'
# moive_url_list = []

# request = urllib.urlopen(url)
# response = request.read()
# response = unicode(response,'GBK').encode('utf-8')

# soup = BeautifulSoup(response,'html.parser')

# a  = soup.find_all(href = re.compile("/html/gndy/jddy/\d{8}/\d{5}.html$"))
# for i in a:
# 	pattern = re.compile(r'^\d{4}')
# 	string = i.string
# 	match = pattern.match(string)
# 	if match:
# 		moive_url = url + i['href']
# 		moive_url_list.append(moive_url)


# for j in moive_url_list:
# 	req = urllib.urlopen(j)
# 	res = req.read()
# 	try:
# 		res = unicode(res,'GBK').encode('utf-8')
# 	except UnicodeDecodeError:
# 		continue
	
# 	soup =  BeautifulSoup(res,'html.parser')
# 	a_tag = soup.find_all('a')
# 	for m in a_tag:
# 		pa = re.compile("^ftp://ygdy\d{1}:ygdy\d{1}@y\d{3}.dydytt.net")
# 		st = m['href']
# 		ma = pa.match(st)
# 		if ma:
# 			print st
#上面代码虽说能找出电影连接，但是不是最新的。
################################################################			

from bs4 import BeautifulSoup 
import  urllib
import  re

url= 'http://www.ygdy8.net/html/gndy/dyzz/index.html'
moive_url_list = []
moive_name_list = []

request = urllib.urlopen(url)
response = request.read()
response = unicode(response,'GBK').encode('utf-8')
soup = BeautifulSoup(response,'html.parser')

a = soup.find_all('a',class_="ulink")
for  i  in  a:
	moive_open_url = 'http://www.ygdy8.net' + i['href']
	req =  urllib.urlopen(moive_open_url)
	res = req.read()
	try:
		res = unicode(res,'GBK').encode('utf-8')
	except UnicodeDecodeError:
		continue
	so = BeautifulSoup(res,'html.parser')
	a_tag = so.find_all('a')
	for j in a_tag:
		pattern = re.compile("^ftp://ygdy\d{1}:ygdy\d{1}@y\d{3}.dydytt.net")
		string= j['href']
		match = pattern.match(string)
		if match:
			moive_url_list.append(string)


for m in moive_url_list:
	print m

