I am trying to implement the webrequest get and post methods. I am accessing a secure site that requires login authentication and was told that instead of sending a login request all the time; login to the site and capture the cookie and use this to send it in the header to get results from the page. In other words let us use google for example.
Lets say google required u to login before searching for something. Now I login through my browser and leave the site logged in. Now I send a httwebrequest from my program that includes the cookie authentication details in the request header and get results for say ?param=sports. Now when I increment the page number like &page=3, I am still only getting page 1 results. Here is some code:
My boss has given me an asp.net web application that was done by his former employee. Now my boss has no contact with that employee. He wants me to make some changes in that web application. I went through the application. There is no documentation present for that application . Its a huge application. It contains many directories. I think it was a team work. It has many asp pages, aspx pages, jsp pages, xml pages, css pages, etc... How to understand other person or team's website or web application that was written in asp or asp.net and vb or vb.net?
I have a Search web reference (from a Search Server 2010 express install) in a vb.net application that is utilizing the QueryService Class to search a production Sharepoint foundation 2010 site.At a previous point in time, we had created a proof of concept on an entirely test system that has since been turfed. From my recollection on this test system when documents were uploaded as a specific site content type (that inherits from document) and metadata was provided, we could search for specific metadata by making managed properties for each, and search results would be returned as documents (with the isdocument flag set to true). Viewing the document then became simple, as we could simply use the filename and path to display the stored file.Now we are developing a production system and we have encountered a new behavior, where these results now are returned as aspx results such as
[URL]
This of course makes it terribly difficult to locate and view the document, we can extract the Title which will then give us the name of the file with no extension, but that hardly helps, as the FileExtension data is aspx, not the documents file extension, so we don't have a full filename. We could display the page returned as a result, but would much prefer the document itself.
I've made a test document library, with just bare bones setup, (not using the site content type, or site columns) and uploaded some documents on the same site, and they are returned in the same fashion, so I don't believe the document library, or content type are the issue.
With a fairly limited understanding of both Sharepoint and Search Server, I don't know if this is a setup issue with the search service itself, with the site configuration, or with the querypacket I am sending. We also have a third party application (Knowledgelake) installed on the server that ties into sharepoint which could have changed configuration somewhere as well?
I don't think the query packet has changed since it was working in the proof of concept, other than the custom data column names. I will provide it here in case there is something glaringly obvious to an external reader.
I am using Stanav's PdfManipulation2.AddImageToPage(..., Pages) for iTextSharp library (v5.1.3) and was not able to add the image to specific pages: e.g. to add to only page 1 and 3 of a five pages PDF
Dim Pages() As Integer Pages = New Integer() {1, 3} PdfManipulation2.AddImageToPage(..., Pages)
Calling AddImageToPage() with the the optional Pages parameter will add the image to all pages though.
I have a form on a .NET webserver using aspx pages with the vb code-behind pages. The form collects the entered data and then emails it, that part works fine.Right after it gets emailed, this code checks to see if the email was sent or not and then updates the landing page accordingly. I used to use just an IF THEN statement, but now that doesn't work in these new .NET pages. Here's the end of the code that I am having difficulties with:
I have heard of setting server specs to have HTML pages parsed as PHP pages by changing the .htaccess or httpd.conf files. Is there a similar solution to have HTML pages parsed as ASPX pages?
A website I am working on does not want their PageRank to lower since their HTML pages are already indexed and used on other websites but wants to use some dynamic features of aspx pages (like include the navigation file so that changes to the nav will only have to be done in that one file).
Im trying to create a software which can post into yahoogroups.I've done trying to log in yahoomail but my problem is when I am going to post in yahoogroups, I am turning back in to yahoo log in page.
I have an ASP.NET web application (VB.NET) using an Oracle database. On an insert, I need to get the identity of the inserted row back. I am trying to use RETURNING, but I keep getting a value of 1 returned.
Dim strInsert As String = "INSERT INTO L.TRANSACTIONS (LOCATION_KEY, TRANS_CREATOR, TRANS_EMAIL, TRANS_PHONE) VALUES (:location_key, :trans_creator, :trans_email, :trans_phone) RETURNING TRANS_ID INTO :ukey"
My brain may just be fried since it's the end of the semester, but my teacher kicked this back to fix and resubmit...she said that I am "returning the name of your function rather than your variable name".
Here is my code for all 3 forms.
code:
These are forms to calculate the cost of getting a cell phone package.
i have this program which records the every printing made on a local computer.I am to record the number of pages of the document to be printed and the number of printed pages. i have this code but i don't know the number of pages and number printed pages always return 0 [Code].
I have a series of WebRequests in my app all formatted like this
Try request.Timeout = 30000 ' Get the response and read it to the end [HTML of a page]
[Code]....
After roughly 250 iterations of about 5-6 WebRequest per iteration, the app hangs and starts hitting the WebException "The remote name could not be resolved: 'hostnamehere' ". WebPages also don't load through my browser when this happens, I have to close the app for my internet to work again. This app is multithreaded and I was testing with 100 threads (in a threadpool, MaxThreadCount set to 20, addressof the sub which performs the WebRequests). A test with 20 threads yielded identical results - Same WebException also right around 250 iterations processing the same sub. I have tried setting the ConnectionLimit of the request to the thread count with no luck.
I am trying to use httpwebrequest but I keep getting a message that says "Create is not a member of webrequest" every time I try and use the msn example code.
vb Dim myReq As HttpWebRequest = WebRequest.Create("http://www.contoso.com/")
I dont know what i did wrong Here is the code:Im trying to login to uploadee btw, a paid per download site.
Dim Postdata As String = "sec_login=sec_user_in&email_two=" & TextBox1.Text & "%40hotmail.com" "&password_two="& TextBox2.Text & "&x=0&y=0" Dim tempcookies As CookieContainer
[code].....
Variable 'encoding' is used before it has been assigned a value. A null reference exception could result at runtime.Variable 'tempcookies' is used before it has been assigned a value. A null reference exception could result at runtime.
I would like to search an online (internet) database and import the result in my own Windows program. As I understand from reading and searching I first have to do a get from the default (search) page to get the cookies for instance and then do a post providing the filled input fields and adding the previously collected cookies. So far getting the information seems to work but submitting the data back to get the actual search result doesn't seem to work. This is what I have so far: [code]Anybody that can point me in the right direction? I my assumption correct that I first have to do a Get and then Post the correct data back?
I think this is a simple question for you, but I don't understand other cases of webRequests, so I asked here: How can I make this webRequest asynchronous? Dim sBuffer As String Dim oRequest As WebRequest = WebRequest.Create(url) oRequest.Method = "GET" Dim oResponse As WebResponse = oRequest.GetResponse() Dim oStream As New StreamReader(oResponse.GetResponseStream()) [Code] .....
Does anyone at least have some related links on fetch requests and scraping?I am trying to implement the webrequest get and post methods. I am accessing a secure site that requires login authentication and was told that instead of sending a login request all the time; login to the site and capture the cookie and use this to send it in the header to get results from the page. In other words let us use google for example. Lets say google required u to login before searching for something. Now I login through my browser and leave the site logged in. Now I send a httwebrequest from my program that includes the cookie authentication details in the request header and get results for say ?param=sports. Now when I increment the page number like &page=3, I am still only getting page 1 results.[code]
The current version of my application was using the WebClient object to download files from the internet and all was working well until some users reported that they were getting server timeouts. I looked into this issue and decided to move to the WebRequest object as it allows me to specify a Timeout value. The problem is, when using it synchronously, if I specify a timeout value of -1, I am worried that in the event of some weird server anamoly, my application would hang as it would be waiting for a response that isn't coming. So then I thought of using the WebRequest ansynchronously.
I am still learning about Get / Post statements in VB, and wondering if you have to call the Get commands to images, and what not, or just the main Post / Get statement and the rest will auto-load?
I've got an problem with my GUI. My program has to request a couple of websites (POST) and then store the responses in local files.While requesting the sites, a GUI textbox (tbLog) should inform the user, what site actually is being requested and a "STOP" butten should enable the user to stop the whole process.Now, when I start work, the GUI freezes and the user cannot do anything until all websites are done. There is no update of tbLog.text at all and no stop button can be pressed.Me.Refresh() does not help, to.[code]
I have a loop with WebRequest objects which I declare every iteration. I thought it would be more efficient to declare them outside of the loop and only modify the one parameter that changed but it didn�t work. Why is that? It seems like the loop acts like a subroutine or other block that kills it�s objects at the end of each iteration. Is this the case? It seems other global objects like a text box can be modified this way so why not the WebRequest?
I want to write in a csv file that is in an url, but it`s not work look my code:how do i do to write in a file?Dim request2 As Net.WebRequest = Net.WebRequest.Create("http://www.bbb.com/teste_download/dados.csv")
I am writing some rest calls to the LinkedIn API using web request and have hit an encoding issue.The LinkedIn API takes requires an encoded URI as one of the parameters on the lookup i am trying to run.
We are calling the api through .Net's WebRequest.Create; and when we pass in the URI it decodes it and as a result we can't get the data from the call. Im trying to open a URI like this[URL]..Is there a mechanism in the webrequest class to stop it decoding the passed in URI; or a work around for this?
I am trying to "behind the scenes" log myself into a website, from the VB code behind my ASP.NET website. But I am dumbfounded as to how to do this.
As far as I know I should be using the WebRequest or Webclient class. That is about as much as I know. I am not sure how to use the class.
I want to click a button on my website and have its Click event send a username and password to another website. This other site isot affiliated with mine. I realize the concept may seem stupid, but I plan on taking this further later, but Just need to know this now.
I made a small console program to login into a website.After figuring out on how to successfully login to the website i ran into a other problem... that problem is visiting the inbox page.The problem is as follow: when i visit the page i get logged out :sMy problem is i think that i don't know how to pass the cookie/session correctly since this is not a post/submit action...When i debug the site and click on the hyperlink i only get get response.
I have a web request that can take 30-90 seconds to complete in some cases(most of the time it completed in 2-3). Currently, the software looks like it has hung if the request takes this long. I was thinking that I could use background worker to process the webrequest in a separate thread. However, The software must wait on the request to process before continuing. I know how to setup the background worker. What I am unsure about is how to handle waiting on the request to process.Do I need to create a timer to check for the results until the request times out or is processed?
Looking for the best way to determine if a URI exists in VB.NET without downloading the actual content. System.IO.FileExists/My.Computer.FileSystem.FileExists can be used locally to determine if a file exists, is there an equivalent for the Web?
Currently I am using a HttpWebRequest to check URI existance using the ResponseStream. This populates the stream if the target does exist and throws an exception if it doesn't. The function is being expanded to also check for PDF files (typically 5MB +), images, etc and it will be a waste of time/bandwidth to actually populate the content into a stream.
In the case of "Success" (the target does exist) I do not wish to download the file or page, simply to end up with a Boolean which indicates the whether something exists at the end of this URI.