I have been reading books trying to learn .net on my own. I am working on an exercise that asks me to create a 1MB file. what i have below does work, but it seems to me there would be a more efficeient way. It seems like what i am doing would cause a lot of overhead. making a 1MB file easier than what i have?[code]...
I've an arraylist having 30000 items in it, what's the best way of creating a text file on the fly from an ASP.NEt page? Currently I'm using the code below but it times out with large data,[code]
I'm trying to create an application that reads in a log file, then converts each line to a class(I'm databinding a datagrid to the finished ObservableCollection).The implementation I've done but I'm struggling to reduce the time it takes to itterate the lines and convert them to classes.small log files are fine its the larger log files (25k lines +) which is my concern. I'm wanting to do this with say a max delay of around 5secs from loading the text file.my current implemenation of a log file I have of size 85k lines is several minutes which isn't great.
We have some information that we need to write (about 18KB) to a .txt file stored in one of our network drives. The file is re-written about once every 15 minutes but is read practically at least every second. We are currently using StreamWriter to write file.The file server is in remote location and the round trip ping varies from <1ms to 15ms.
The problem is, sometimes it takes as long as 6 seconds to write the contents to the file, which is definitely way too long even after we take consideration of the network speed. therefore, I am just wondering if there is any efficient way to write the file using VB.Net to improve the performance? Java has a very good tool named BufferedOutputStream, which unfortunately is not available in VB.Net (or I just have not found it).
I am looking for some guidance with the most optimal or efficient way to export a dataset to an excel file. I have searched the forums and found a number of methods to do this. However, the code I am currently using is very inefficient as it is itterating through the rows and columns to create the excel file. The app I am working on needs to be able to create 50 or so excel files each with 500+ rows and 40-50 columns. This method seems to inefficient for my use.
I have seen some code out there to use an ODBC call to excel to create the file. I am wondering if someone can verify that this is the most efficient way to meet my goal. It should also be noted that after the export I will want to do some basic formatting of the sheet. Adding bold column headers, autofilters, and autofit all columns.
It seems that when using With blocks in VB.NET, the resulting MSIL larger than w/o. So this leads me to ask, are With Blocks really more efficient? MSIL is what gets JITed into native machine code, so smaller code size should imply more efficient code, right?
I need to verify if a certain user exist on my asp.net site.I want to know if which of these two functions is more efficient, faster and better compared to each other and why.
I am working on an application that will need to read tons of records (close to 500,000) from one table and insert them into another set of tables in the same database. I though about using SSIS package for this but our DBAs don't want to use that. Now, I am thinking of a multi-threaded approach. I am thinking that I can have a few thread started that will read say (500) records at a time and insert them, then come back and read more.
Now, say I spawn off 3 threads of this application. The first thread reads 500 rows and starts processing them. Can I lock these rows that were already read so that the next thread does not pick them up? I am trying to find some articles regarding this on the internet, but perhaps I am not searching for the correct terms in Google.
I'd like to sample the data from the Arduino @ roughly 200Hz. Is there a more efficient way to do this than using a timer (which is what I am currently using)? Currently I am using[code]....
if there is no need for my variable to store negative values, is it more efficient to use UInteger instead of Integer? both are 4bytes and i've got a feeling unsigned works faster though i'm nt sure about it. note that the range of my variable is 0 to 2147483647, hence i could use a UInt32 or Int32 and both would give me the same result so i'm wondering about optimum performance.
I have a program that must initialize a table that is usually about 10-15 thousand records. I have attempted several means to empty the records from it, but it always takes several minutes to accomplish the task or in some cases like a 'bindingsource.clear' command, doesn't work at all. I currently have it working using a simple loop to go through each record at a time and 'removecurrent'. but this seems extremely inefficient and it is quite slow.The real time consumer seems to come in with the tableadapter.update and tableadapter.fill that happens after the loop.[code]
I am looping through my dataset to save my database, I thought this might not be the best way. Is there a more resource efficient of doing it?
For i As Integer = 0 To CO.Tables("CO_Del").Rows.Count - 1 'CO is my DataSet dbCommand.Parameters.Clear() AddInParameter(dbCommand, "@IsDelete", DbType.Boolean, 0, DBNull.Value)[code].......
I have a Page class and a PageCollection class in a 3d party ORM framework. I can fill the PageCollection based on parameters (pageid, parentid, url etc..) (SQL query). But I need the data multiple times around the ASP.NET MVC website (Sitemap, Authentication), so I chose to load all pages 1 time and reference that (global) collection.
GlobalClass.Pages //is PageCollection containing all pages
I have now created functions which return a temporary subcollection or single entity based on the parameters mentioned before (pageid, parentid, url etc..).
GlobalClass.Pages.GetByPageId(id) //returns single page entity GlobalClass.Pages.GetByParentId(parentid) //returns subcollection
The site however got very slow.
cache the subcollections (GetByParent())? create internal hash-lookup tables for the collection? Something else...?
Namespace BLL Public Class PageCollection Inherits CustomCollectionBase
I have an ASP DataGrid and I'm applying sorting to it. Well, as I was looking at an example, they had a function similar in function, different in name, to[code]...
While learning to Orientate of the Last layers of a rubik's cube, I thought that it would be fun to make a vb.net application for it(I know that it's super hard to do)I'm starting around the center cube, but was wondering if I did a part right..I'm working with squares, and was wondering the length from one corner to another. [code] Trying to figure out the distance from the top left to bottom right,I used the distance formula to test it out and got 70.71067811865475 sqrt((-50-0)^2+(-50-0)^2)Is there an easier or more efficient way to get this number than using the distance formula?
I've been using LINQ so much in the last couple of weeks that when I had to write a one line function to remove < and > from a string, I found that I had written it as a LINQ query: [code] My question is, is it better to do it with LINQ as above or with StringBuilder as I've always done, as below: [code] Both work, the second one is easier to read, but the first one is designed for executing queries against arrays and other enumerables.
I made this small location class so that I can better handle address information which I am going to be sending off in requests to the Google Maps API. One thing I have left to do is some validation to make sure that the address has enough information to return a result back. For the application, the level of accuracy should be as loose as a single City (meaning that it should work as long as a zip code or city/state is provided since it will find the geographic center of that area automatically in Google Maps)
I was wondering what was the best way to search a database. Is it possible to perform bubble sort etc or any sort of binary search to the database instead of doing a linear search. If so how would I go about doing this through an Sql statement.
Ok so I have this program that shows data in a treeview and that data is from a datatable.I'm basically looping through the datatable row by row and adding nodes.The data looks something like the following:
[code]...
I have a thread that is hitting the Database (where these records are stored) every 5 seconds to see if the list is smaller,bigger or changed.If so, I was just recalling the load_form sub but this was taking too long as the lists could potentially be very large. Also, there is a lot of logic that comes into play depending on what user is displaying the records in the treeview.I was wanting to know what would be the fastest way to compare the datatable that was previously used to initially create the nodes in the treeview to a tempdatatable that has the currently updated/added/removed rows and update the treeview accordingly?The only thing I could come up with was basically looping through the tempdatatable and comparing each row to the rows in the original datatable and plucking out the ones that had no match.Then recursively looping through the nodes in the treeview to find the spot that the new node should be added (based on that datarow's values).Or is there a better way to do this and I should scrap this?
I have written a program that goes to user defined webpage and grabs images of certain ids from the site by going through the source code of the site. My current program uses a webbrowser but what I am curious of is, if I write code that grabs the html document of the site (i assume thats just the text and doesn't download the whole page), then go through the text looking for the element id, then grab the image, would this more be efficient?
Efficient meaning, would it do the same task using less data, which would decrease the load time of the site. Also is it also even possible to go through an html documents element and grab a certain link? I know how to grab the html document, it was the grabbing of the link that I got stuck at.
url...But what I'm looking for is just see if any alpha (0-254) is used in at least one of the pixels. Would the Bitmap.Lockbits and BitmapData method be the most efficient (speedy?) or is there another way. I've tried toying around with ColorPalette and the Alpha flag, but never could figure it out (and even if it is a way to check to see if the alpha channel is used at all).
Whats the best way to handle the connection to a db when it's accessed often? I have an app where the user selects an item from a drop down list(call to the db) and then a picture of the item is returned along with it's relevant information(another call to the db). Small application but basically every action results in a call to the database. Should i just be filling like a global dataset and then accessing the information from there(if thats possible)? The database is stored locally(currently an Access Db but will be SQL in the future)
I want to create a simple program which can convert Units of Measure (e.g inches into centimeters, milliliters, meters, feet, etc). I want to create a function which has two parameters (unit to convert from and unit to convert to) and returns a conversion value (e.g. funciton passed parameters values inch and centimeter and returns value 2.54). In terms of strategy for the least processing overhead I am exploring a few options. I guess the slowest way would be some kind of nested If/Then or Case loop, so I wanted to use something similar to a 2D table lookup in excel (where the top row and the left column (key?) would have the unit types, and the conversion values would occupy the cells in the table body). Is this possible in VB? Could I load all of this into a 2D array or something? I read a bit about hashtables and dictionaries as well, Ideally Im looking for the least processor-hungry code since this function may be called upon frequently. Ill be using other functions to perform similar lookups on other 2D tables elsewhere too (i.e. matching a value on the top row and the left column and returning the coincident cell value in the table body).
Just wondering what the most efficient way to remove multiple characters that are next to each other is.so that running would become runing - notice letters can appear in there multiple times - the n in this case.[code]
The code below shows me (I think) that the "for each" loop is about 10% faster than the "i to n" loop, but the "for each" loop creates 567k in new memory? Is this right? Which way is generally most efficient with regards to speed and memory usage? If you want to run this code in VB just add a button and 2 labels to a form.