I am trying to unshorten urls and have not been able to find code (vb.net/c#) to do this. These are the twitter shortened urls and I guess I could try and access one of the web services available and do a httpwebrequest but would prefer to find some programmatic way of doing this.
This in ref to this question that i asked but never got answered ASP.NET 4 ACCESS DATA TO APPLY TO NavigateUrl but suppose thats redundant now.
I need to generate dynamic URLs from a database in code behind and then use them in a list view but i cannot find anywhere on the internet or in my book that covers something like this.
Im getting data out using below in a code behind page:
'portfolio navigation data Dim rdrPortfolioNav As SqlDataReader Dim cmdPortfolioNav As SqlCommand = New SqlCommand()
[Code]....
how to loop through each record on the code behind page to generate all the urls then how to show these on the actual page within the list view.
I have a list of 100,000 urls in list(Of string) which can contain urls in the form. [URL] i have tried using a combination of regex and the Uri class, but that didn't help, so i dumped the code. How do i filter these duplicates and keep just one of these url
I need to find .MP3 format URLS in a HTML source code.So how could i do that?Lets say i have:
Dim wcClient As New System.Net.WebClient Dim data As System.IO.Stream = wcClient.OpenRead(inbox.ToString) Dim reader As System.IO.StreamReader = New System.IO.StreamReader(data) reader = reader.ReadToEnd() reader.Close()
so how could i find all the .MP3 urls which are in the source code?
I've found some examples using RegEx but im not really sure how to use the RegEx pattern to find MP3 urls in the source code.
extracting all URL's from a .PDF File and load them in a listbox using itextsharp. After searching i found the following code from STANAV but don't it gives an error on the line of code in red and how do i load this URL links in a listbox.
I'd like to know if there are other ways to write a submission program (submitting urls or pad xml to directories, submission sites, blogs) other than using the webBrowser. I've evaluated some shareware programs that submit urls or data very quickly, like 100s in a matter of mins. I couldn't write a program using a webBrowser because it might take 15 secs just to load the url. How are they writing the code??
Possible Duplicate: What's the term for the part of the URL after the question mark?
I have a company Website in Visual Studio / VB / ASP.NET 4.0. I noticed a competitor of ours also has .aspx extensions in their website. They have strings at the end of their website URL. For instance, when you click a language, it has this after the dot com:
/main.aspx?langtype=2057&locale=en-be
And of course it goes to that language/culture and the "be" is, I assume, the uiculture/subculture.But how do you get these strings in your own URL?
I'm trying to write out animals from a database. All the samples I see for creating View pages have hard-coded "nodes" in the URLs. I can get my first view to work but my second one doesn't write out the correct URL to go to the third one. So, for example, I want my URLs to be:
... I'm guessing I'm missing something in a route path but what is it? It can't be that I have to know the path I'm at on every page in order to write out the URL - can it?
I want to extract all the URLs of favorites. Mine code is as below- Code is adding all the Items into Listview. I want that on ListView_Click,dat site is opened.
I am pulling urls into excel, but excel formats the data. if a website displays 000.4500 excel formats it to .45.... i need everything exactly like displayed on the webpage, so excel should not do anything with it.
I tried formatting the worksheet to text (keep everything as entered exactly) and that did not solve the problem.
I can show code, but i'm not sure if this is a coding problem at all. I am just confused and can't seem to find any good information on this problem.
I somehow managed to make my company's website in 9 languages, using Visual Studio 2010 / VB / ASP.NET 4.0. I believe I'm using sessions. But you can tell for sure if you see the code provided below. I know this message is long, but I really need help.
It's a multilingual site, and I managed to put flags on the homepage. When I click a flag, the page's text changes to that language. When you click the French flag, it gets the information from the "FR" .resx resource file in my apps_GlobalResources folder. It stays on that language for the end-user's entire session. Great! Well, not so great.
The URL, for instance, about.aspx, remains about.aspx. Granted, the text changes to French, but I've been told that it's recommended to make it look like, if the client chooses French, for example, [URL] (For reference, if anybody is kind enough to delve into this issue, which may be a simple one, all of my files I've used in this are included below -- and they're kinda lengthy)
I'm trying to make some friendlyurls in my vb.net (.net 4) project and I'm trying to do it using something I read about global.asax and Application_Beginrequest but I can't get it to compile.
Before I start, you'll probably tell me to search the forum to get the answer.But really I am searching net + this forum from last 5 days to get solution for this problem but unable to find..So here I start, probably not a new problem.
Asp.net has turned out to be alot easier to use than PHP (so far). However, I have been searching for a while and simply cannot figure this out. How do I get the variables that are contained in the url of my page (that originate from a form that had the method "GET") and utilize them? For example, my page would be [URL]. How would I get the value of some variable?
I am trying to capture URLs in HTML file which appears like
<a[string]href[space(s) or nothing]=[space(s) or nothing]["][url]["][string]>
I found this code but it does not work well.
Imports System.Text.RegularExpressions Public Class Form1 Private Sub Button1_Click(ByVal sender As System.Object, ByVal e As System.EventArgs) Handles Button1.Click Dim rx As New Regex("[<]a[s][wW]*[href=](?<word>S*)[sWw]*[>]", _ RegexOptions.Compiled Or RegexOptions.IgnoreCase) Dim text As String = "<a href=http:// name=as>" Dim matches As MatchCollection = rx.Matches(text) For Each m As Match In matches MsgBox(m.Groups("word").Value) Next End Sub End Class
I am iterating through the lines of a RTB that has captured the HTML of a website. I want to check each line for a URL (just the first one is fine) ---- I can create a substring when it finds an http:// but I cannot figure out how to get rid of everything after .com or .org, etc.I have found a regex that supposedly does it but am not sure how to implement it.... here is what I have so far: For Each currentLine As String In rtb1
I have a list of URLs in a listbox and I want to remove the Urls containing some unwanted domain for example google.com also I want to know if is possible remove all URLs no containing a domain for example yahoo.com
I have a list of 100,000 urls in list(Of string) which can contain urls in the form.[URL]..i have tried using a combination of regex and the Uri class, but that didn't help, so i dumped the code. i also tried using this code, but it will only remove duplicatse of exact form, since its not domain specific.
list = new ArrayList<T>(new HashSet<T>(list))
How filter these duplicates and keep just one of these url if it contains the same name e.g yahoo.
all URL are of different domains, but can usually have duplicates like the example i gave above
I've a situation where I am trying to extract a large number of URLs from a site, but the site which provides the URLs has an obstacle. Rather than the link directly to the site, they provide a link to an internal page, which automatically redirects me to the URL I need.
i.e. <a href="www.StartingSite.com/outgoing/1234" ...> Example.com</a>
which passes me to an internal page, then automatically routes me to Example.com
I've suspected if I can do this, it would be through cookies.
Perhaps a little more work, I could create a page which calls each page, then I could acquire the URLs from my history, but my browser would crash for the # of URLs I'm extracting.
any means I could gather the end URL of this link?