锘??xml version="1.0" encoding="utf-8" standalone="yes"?>亚洲综合久久久,人妻无码久久一区二区三区免费,99久久无色码中文字幕http://www.shnenglu.com/gaimor/category/20048.html涓氫綑CPP涓撳zh-cnSun, 08 Dec 2013 00:35:58 GMTSun, 08 Dec 2013 00:35:58 GMT60C璇█璋冪敤python鑴氭湰3http://www.shnenglu.com/gaimor/archive/2013/12/07/204649.htmlccsdu2009ccsdu2009Sat, 07 Dec 2013 07:48:00 GMThttp://www.shnenglu.com/gaimor/archive/2013/12/07/204649.htmlhttp://www.shnenglu.com/gaimor/comments/204649.htmlhttp://www.shnenglu.com/gaimor/archive/2013/12/07/204649.html#Feedback0http://www.shnenglu.com/gaimor/comments/commentRss/204649.htmlhttp://www.shnenglu.com/gaimor/services/trackbacks/204649.html
from bs4 import BeautifulSoup

def list_get(file):
    soup 
= BeautifulSoup(open(file))
    alist 
= soup.find_all('a',class_ = 'link')
    list 
= []
    
for i in alist:
        list.append(i.
get('href'));
    #
for i in list:
    #    print(i)
    
return list

if __name__=="__main__":
    list_get(
'List.htm')

list_get鍑芥暟榪斿洖鐨勬槸list瀛楃涓插璞?br />鍏禼璇█璋冪敤鐨勪唬鐮佸涓?
#include <stdio.h>
#include 
<stdlib.h>
#include 
<Python.h>

char* GDALPythonObjectToCStr(PyObject* pyObject);

int main(int argc, char *argv[])
{
    Py_Initialize();  
    
if(!Py_IsInitialized())   
    {  
        
return -1;  
    }  
    
    PyRun_SimpleString(
"import sys");
    PyRun_SimpleString(
"sys.path.append('./script')");
    PyObject
* pModule;
    PyObject
* pDict;
    PyObject
* pFunc;

    pModule 
= PyImport_ImportModule("list");
    
if(!pModule)
    {
        printf(
"can't find list.py");
        system(
"PAUSE"); 
        getchar();
        
return -1;
    }
    
    pDict 
= PyModule_GetDict(pModule);
    
if(!pDict)
    {
        
return -1;
    }
    
    pFunc 
= PyDict_GetItemString(pDict,"list_get");
    
if(!pFunc || !PyCallable_Check(pFunc))
    {
        printf(
"can't find function [list_get]");
        getchar();
        
return -1;
    }
    
    PyObject
* args = PyTuple_New(1);
    PyTuple_SetItem(args,
0,Py_BuildValue("s","List.htm"));
    PyObject
* value = PyObject_CallObject(pFunc,args);
    
int ret = PySequence_Check(value);
    printf(
"check:%d\n",ret);
    
int length = PySequence_Size(value);
    printf(
"length:%d\n",length);
    
    
int i = 0;
    
for(;i<length;i++)
    {
        PyObject
* obj = PySequence_GetItem(value,i);  
        
//char* str = PyBytes_AS_STRING(obj);
        char* str = GDALPythonObjectToCStr(obj);
        printf(
"link:%s\n",str);    
        free(str);
    }
    
    Py_DECREF(args);
    Py_DECREF(pModule);
    Py_Finalize();    
    system(
"PAUSE");    
    
return 0;
}

/* Return a NULL terminated c String from a PyObject */
/* Result must be freed with GDALPythonFreeCStr */
char* GDALPythonObjectToCStr(PyObject* pyObject)
{
#if PY_VERSION_HEX >= 0x03000000
    
if(PyUnicode_Check(pyObject))
    {
        
char *pszStr;
        
char *pszNewStr;
        Py_ssize_t nLen;
        PyObject
* pyUTF8Str = PyUnicode_AsUTF8String(pyObject);
        PyBytes_AsStringAndSize(pyUTF8Str,
&pszStr,&nLen);
        pszNewStr 
= (char*)malloc(nLen+1);
        memcpy(pszNewStr,pszStr,nLen
+1);
        Py_XDECREF(pyUTF8Str);
        
return pszNewStr;
    }
    
else if(PyBytes_Check(pyObject))
    {
        
char *pszStr;
        
char *pszNewStr;
        Py_ssize_t nLen;
        PyBytes_AsStringAndSize(pyObject,
&pszStr,&nLen);
        pszNewStr 
= (char*)malloc(nLen+1);
        memcpy(pszNewStr,pszStr,nLen
+1);
        
return pszNewStr;
    }
    
else
    {
        
char *pszStr = (char*)malloc(1);
        pszStr[
0= '\0';
        
return pszStr;
    }
#else
    
return PyString_AsString(pyObject);
#endif
}


ccsdu2009 2013-12-07 15:48 鍙戣〃璇勮
]]>
浣跨敤Beautiful Soup 瑙f瀽html鏂囨。http://www.shnenglu.com/gaimor/archive/2013/12/07/204645.htmlccsdu2009ccsdu2009Sat, 07 Dec 2013 03:17:00 GMThttp://www.shnenglu.com/gaimor/archive/2013/12/07/204645.htmlhttp://www.shnenglu.com/gaimor/comments/204645.htmlhttp://www.shnenglu.com/gaimor/archive/2013/12/07/204645.html#Feedback0http://www.shnenglu.com/gaimor/comments/commentRss/204645.htmlhttp://www.shnenglu.com/gaimor/services/trackbacks/204645.html灝辨嬁瑙f瀽html鏉ヨ錛屼嬌鐢?Beautiful Soup 灝辨瘮浣跨敤libtidy鏂逛究寰堝 - 褰撶劧涔熸湁鍙兘鏄?div>Beautiful Soup灝佽鐨勫緢鍘夊鍚?br />
浣跨敤Beautiful Soup鐨勪竴涓緥瀛愬涓?
from bs4 import BeautifulSoup

soup 
= BeautifulSoup(open('List.htm'))
for a in soup.find_all('a',class_ = 'link'):
    
print (a.get('href'))
鐩殑鏄壘鍑篽tml涓璫lass灞炴т負link鐨刟鑺傜偣瀵瑰簲鐨刪ref灞炴у瓧絎︿覆
濡傛灉鏄嬌鐢╟++ libtidy鐨勮瘽
瀵瑰簲鐨勪唬鐮佸涓?
Bool TIDY_CALL tidyFilterCb(TidyDoc tdoc,TidyReportLevel lvl,uint line,uint col,ctmbstr mssg)
{  
    
return no;
}

void extractContent(TidyNode node,TidyDoc doc);

void parseContent(TidyNode node,TidyDoc doc)
{   
    TidyNode child;
    
for(child = tidyGetChild(node);child;child = tidyGetNext(child))
    {   
        
if(tidyNodeIsA(child))                    
            extractContent(child,doc);
        
else
            parseContent(child,doc);     
    }
}

void extractContent(TidyNode node,TidyDoc doc)
{   
    
if(yes == tidyNodeIsA(node))
    {   
        TidyAttr cls 
= tidyAttrGetCLASS(node);
        
if(cls != NULL)
        {   
            
char* value = (char*)tidyAttrValue(cls);
            
if(!strcmp(value,"link"))
            {   
                TidyAttr href 
= tidyAttrGetHREF(node);   
                
if(href != NULL)
                {
                    
char* link = (char*)tidyAttrValue(href);
                    printf(
"link:%s\n",link);
                    
return;
                }
            }
        } 
    }   
    parseContent(node,doc);
}

void tidyParseHtml(char* file)
{
    TidyDoc doc 
= tidyCreate();
    tidySetReportFilter(doc,tidyFilterCb);
    tidyParseFile(doc,file);
    TidyNode body 
= tidyGetBody(doc);
    TidyNode child;
    
for(child = tidyGetChild(body);child;child = tidyGetNext(child))
    {   
        parseContent(child,doc);
    }    
    tidyRelease(doc); 
}
榪樻槸寰堝暟鍡︾殑

褰撶劧涓嬮潰鐨刾ython浠g爜涔熻兘瀹屾垚浠誨姟:
from bs4 import BeautifulSoup

soup 
= BeautifulSoup(open('List.htm'))
list 
= soup.select('a[class="link"]')
for a in list:
    
if a.has_attr('href'):
        print (a.
get('href'))
濡傛灉鎯沖垎鏋愮綉欏墊垜瑙夊緱BeatifulSoup緇濆鏄竴涓埄鍣?br />閾炬帴:
http://www.crummy.com/software/BeautifulSoup/bs4/doc/


ccsdu2009 2013-12-07 11:17 鍙戣〃璇勮
]]>
久久久久人妻精品一区二区三区| 天天爽天天狠久久久综合麻豆| 久久精品国产亚洲av麻豆色欲| 久久精品无码专区免费东京热| 97久久天天综合色天天综合色hd| 久久这里只有精品久久| 久久国产美女免费观看精品| 一97日本道伊人久久综合影院| 热re99久久精品国99热| 中文字幕成人精品久久不卡| 久久精品国产亚洲AV不卡| 国产99久久久国产精品小说| 国产精品久久久久影院嫩草 | 色婷婷综合久久久中文字幕| 国产99精品久久| 中文字幕亚洲综合久久菠萝蜜| 久久精品天天中文字幕人妻 | 久久无码专区国产精品发布| 久久99精品国产一区二区三区| 久久综合色区| 办公室久久精品| 999久久久免费精品国产| 久久伊人五月丁香狠狠色| 日韩精品久久久久久| 少妇久久久久久被弄高潮| 日日狠狠久久偷偷色综合免费| 久久久久久久综合日本亚洲| 久久久噜噜噜久久中文字幕色伊伊 | 久久91综合国产91久久精品| 中文国产成人精品久久不卡 | 日韩欧美亚洲综合久久| 久久国产福利免费| 观看 国产综合久久久久鬼色 欧美 亚洲 一区二区 | 久久99久久无码毛片一区二区| 色妞色综合久久夜夜| 久久综合亚洲鲁鲁五月天| 久久精品国产一区二区三区不卡| 伊人久久综在合线亚洲2019| 久久国产精品99精品国产987| 久久精品国产清高在天天线| 久久国产色AV免费看|