Hi, I'm Peter and a new member here.
Who can help me out?
I need to extract all the urls from a simple text file into a new file, every url in a new line.
The urls may have all different formats...
(Is this possible with the free version of textcrawler?)
thx in advance
How to extract all URLs from a text file?
- DigitalVolcano
- Site Admin
- Posts: 1863
- Joined: Thu Jun 09, 2011 10:04 am
Re: How to extract all URLs from a text file?
You need to use a regular expression to find urls.
Such as:
(This may not be perfect for your needs)
You can then click extract. Unfortunately the Extract function is only in the Pro version.
Such as:
Code: Select all
\b((ftp|https?)://[-\w]+(\.\w[-\w]*)+|(?:[a-z0-9](?:[-a-z0-9]*[a-z0-9])?\.)+(?: com\b|edu\b|biz\b|gov\b|in(?:t|fo)\b|mil\b|net\b|org\b|[a-z][a-z]\b))(\:\d+)?(/[^.!,?;"'<>()\[\]{}\s\x7F-\xFF]*(?:[.!,?]+[^.!,?;"'<>()\[\]{}\s\x7F-\xFF]+)*)?
You can then click extract. Unfortunately the Extract function is only in the Pro version.