I have a VB.NET app that has DataGridViews and Drop Down boxes.In VB6, the FlexGrid was fast, scrolling was smooth.Now, in VB.NET, the DataGridView is really slow to scroll through, and the dropdown boxes are sluggish as well.For the interesting bit. If I run Microsoft NetMeeting and either leave it open or click the X to close it (it still stays as a resident program in the system tray though), all the data grids and drop downs work smoothly again.If I actually terminate NetMeeting, they go back to slow/sluggish again.What is it (was it?) in NetMeeting that maybe can be incorporated into my program (some Microsoft library?) that changes how the graphics redraw?
I'm implementing a poor-man's ORM an a legacy application (.NET 2.0 web app, hand-coded SQL queries). I have two data classes: Customer and Order:
Public Class Customer Public Property CustomerId As Integer Public Property CustomerName As String
[code]....
I'm concerned about the performance of this method, namely if I'm pulling multiple Customer's. For each Customer pulled, I have to hit the database again (opening another connection), and getting the orders for that given customer. If I'm not mistaking, it wouldn't take many records to accumulate a large number of open connections.
How can I determine the number of new connections being made and whether or not I'll encounter any connection pooling or other such .NET-imposed limits?
One thought I've had is to pass the SQLConnection object created in the calling class to the getOrdersByCustomerId function and have that function use the (apparently?) already-open connection. I have not tested it namely because I don't know how to determine if it's better than my existing method. Thoughts?
I'm creating a search web service that returns JSON for processing by the client. The service takes a single search parameter, performs multiple lookups on different tables, then returns a custom JSON object. For example, if the user puts in what appears to be a name, I search both the Customers table for a list of the top n customers and the Orders table for the top n orders that have a customer with that name.
i have a datagridview which have 3 columns, two button columns increase or decrease, 3rd displays the value increased or decreased from button. i have datagridview_CellContentClik behind button columns. The problem is increase in value on pressing button is very slow. i mean if i press button quickly the value should increase quickly but its the same slow pace for increasing or decreasing value on pressing button.
I'm using iTextSharp to print a PDF document. Everything goes ok until I have to print the company logo in it. First I noticed that the logo had poor quality, but after testing with several images, I realize that was the iTextSharp rendering it poorly.The test I did to say this was to print the PDF using my code and then edit the document with Acrobat 8.0 and I drew an image. Then printed the two documents and saw the noticeable difference.My question is that if anyone know if this can be due to a scaling problem where I'm failing to tell iTextSharp how it must render the image or is an iTextSharp limitation.
What i really hate about it is if i have a control that custom paints then a transparent control within that control; it paints the transparent control as *button color* until it calls the transparent controls paint event again.... is there anyway to change the behavior of this so that it either paints the last draw over the control until it is actually repainted- or to @ least set a color other than button color?
you think it would @ least use the forms bg color instead - because it looks horrid when you have a black background with all of these near white boxes on it!
i have problem when i generate my program the program takes time till in my datagridview from table adapter in which i already create in a dataset any way how to make the program fill the records faster and decrease the loading time when fill in the datagridview? [code]
I have a question about of capture sound function, someone have a code example for visual basic 2010 in windows 7?, I tried using mcisendstring function but the sound quality, is poor.
I am running windows vista with 1gb memory and 2ghz proccessor. What would be the best FREE version of .NET to use for best performance? Its been years since I have done any PC programming and I would like to do a little bit =]
The upper picture is the output of a Unix task. The lower picture is what I have in VB. Enumerating tasks is not difficult however the rest it is because it's not documented anywhere.
What am I interested in if I want to get 64-bit performance data?
I'm working on a project that calls for high-performance networking with TLS encryption in VB.NET. The stock SSLStream sucks rotten eggs; its 'asynchronous' read/write operations aren't async at all, meaning that they'll operate parallel to each other and to the main thread but that it offers absolutely nothing that resembles async read/read and write/write, which means that reads are totally synchronous to each other and writes are too (only 1 read and only 1 write can be done at a time).
I've done a lot of reading on this failing and it seems like MS has no intention of ever fixing this.Hoping a whole lot not to have to write my own TLS implementation. I've tried the Mentalis open source SSL suite, which will NOT compile. I've tried SocketWrench tools which will not compile under VS 2010 (and whose documentation indicates it's in fact synchronous on its async threads too; can only do 1 async read and 1 async write at a time).
What I want/need is an SSL/TLS implementation that actually works with DotNET that actually allows me to do multiple asynch reads/writes over a TCP/IP transport stream. Anybody have any experience or recommendations on 3rd party toolkits or alternate methods of SSLStream I/O?
I've developed a .NET application that, among other things, does the following:Uses WebClient to retrieve data from a remote server. Serves as a socket server to 2 'satellite' applications run on the same machine or on a LAN.When I run the app in the VS IDE, it works great. It quickly gets the data from the remote server and communicates perfectly with the 2 satellites.However, when I build it and run it as an EXE, the response from the remote server is very slow and its communication with the 2 satellite applications become very poor.Is there some important difference between running an app in the IDE and running it as an EXE that could effect it like this?
I'm overloading a vb.net search procedure which queries a SQL database.One of the older methods i'm using as a comparison uses a Stored Procedure to perform the search and return the query.My new method uses linq.I'm slightly concerned about the performance when using contains queries with linq. I'm looking at equally comparable queries using both methods.Basically having 1 where clause to Here are some profiler results;
Where name = "ber10rrt1" Linq query : 24reads Stored query : 111reads
[code]....
Forgetting for a moment, the indexes (not my database)... The fact of the matter is that both methods are fundamentally performing the same query (albeit the stored procedure does reference a view for [some] of the tables).Is this consitent with other peoples experiance of linq to sql? Also, interestingly enough;Using like "BER10%"
I have this code to ping a list of over 400 switches , the program works fine but for some reason the first time i ping all the switches alot of the attempts time-out(even tho they shouldnt) which makes the program alot slower , if i click ping again less of the attempts time-out(and as a result the program runs faster) and if i click it a third time i usually get a completely accurate result with all the switches pinging back "success" (except for a few which i know for a fact are down anyway) any ideas why this is ?
When writing an If statement, I've always used And when needed like: If 1=1 And 2=2 Then. The only time I ever used AndAlso is if the second condition will error if the first isnt true like: If Not IsDbNull(Value) AndAlso Value=2 Then. However, recently I've heard that AndAlso is better for performance than And as the second condition is only read when the first is true. In this case, should I always just use AndAlso?
These 2 ways of working both work, but I'm wondering if there's a difference in performance:
Dim collection As ItemCollection = CType(CellCollection.Where(Function(i) i.IsPending = True), ItemCollection) For Each item As Item In collection 'Do something here Next
and
For Each item As Item In CellCollection.Where(Function(i) i.IsPending = True) 'Do something here Next
I thought the second one was better as you'd have a variable less and looks cleaner, but on second thought, I'm not quite sure what happens when you put a linq query in the iteration.
I'm sorting a collection of type "Document" (usually around 100k records). Sorting usually takes around 4-5 seconds, and I'm wondering if there's a way to speed up sorting by modifying my "DocumentComparer" class which implements IComparer(Of Document). Since the Compare() method would be called hundreds of thousands of times, are there any performance improvements that could be made here that I've overlooked?
I have this unusual problem with mailing from my app. At first it wasn't working (getting unable to relay error crap) anyways I added the proper authentication and it works. My problem now is, if I try to send around 300 emails (each with a 500k attachment) the app starts hanging around 95% thru the process.
Here is some of my code which is called for each mail to be sent
I have the mission to make a small game for a school project. Pictures boxes, moved by a timer for walking enemies.If there are around 5 or 6 moving picture boxes at the form, my application get troubles and lags.After I kill some enemies (remove them from the Controls Collection of the Form/Panel) It come back smooth.I think the loop of the enemy movement is too complicated but I don't know how to make that simpler.
I'm creating a password hash recovery tool that uses brute force. I have the brute force algorithm run in a background worker. It all works, but some similar apps I tried out have performance ~100x bigger then my app. (mine does 50k keys/second, some others do 5m keys/second or more.)I realize this is partly caused by using a .Net language, but I suspect there are ways to significantly speed up the brute force. A lot of ppl say I should use multi threading, but how is this done in practice? Splitting up the passwords to check would slow down my app I think.
I have a few Oracle procedures that generate/return a large amount of data that I need to write out to a file.I'm currently trying to accomplish with a data-reader. It seems to be working, I've successfully generated a 479mb file without any trouble.It took less than 4 minutes from the time I retrieved the dataReader to complete the file.But the dataReader I get for a particular procedure is crawling.It's unbelievably slow.I modified my code to try and get a better idea of what is going on[code]I'm writing a PL/SQL block that will open that cursor to see if the performance issue exists when I remove the .Net code..
I am dealing with a situation that I need some help with here. I need to improve performance on functionality that records and updates UI with user selection info. What my code current does is
'This is called to update the Database each time the user 'makes a new selection on the UI
Private Sub OnFilterChanged(String newReviewValueToAdd) AddRecentViewToDB(newReviewValueToAdd) UpdateRecentViewsUI() PageReviewGrid.Rebind()'Call Grid Rebind End Sub
So, I've been reading around on the internet and through MSDN trying to find a way to add thousands of items to a combobox without taking 12 seconds.I KNOW, I'm going to incite reflex-responses involving: bad design; use categories and filter; use BeginUpdate ...Here's the thing.I've read through a lot of similiar forum questions as the one I'm asking and none really seemed to answer the question, but these responses are common.The problem is, there is no way to further categorize.I have thousands of customers and each customer can have anywhere from 1 to tens of thousands of designs.The designs have to IDs: DesignID and DesignTitle. The data entry people sometimes have the one and sometimes have the other. If they have the DesignID than all they've got is a number ranging from 1 to (at this time) 60,000. One Customer has 11000 designs. I don't know of any way to further categorize.The reason I use a combobox is for autocomplete. Also, with most customers there are only 10-20 designs, but for some there are thousands.I have to have one practical solution that handles all possibilities.
As far as BeginUpdate or SuspendLayout is concerned, neither have any effect.The sad thing is, the DataSource takes less than a second to load from my DataBase.It's just the setting of the datasource for the combobox that takes so long.It seems to iterate through the entire list once for the bindingsource and then again when I set the DisplayMember.I can set the display member before I set the binding source but that doesn't make any difference in terms of performance.Performance is improved by almost half when I don't use a BindingSource and just set the Combo's DataSource to the List directly, but, sadly, I need to use the bindingSource because there are multiple Combos using the same list.I know WPF uses virtualization for it's combos,.On the internet I found a Technical document on how to implement a Virtual Combobox in C#, but the code was incomplete and I don't really have the time to spend several days creating and debugging something like that.
In the following block of code always returns zero for the CPU usage. However, if I run it through Visual Studio in debugging mode, do a "Quick watch" on the variable "pcCPUCounter", add ".NextValue()" at the top, and tell it to reevaluate, that returns the varying percentage (eg, 5%, 71%, 16%, etc etc) as the processor utilization fluctuates.
Why would the code always print out zero, but the quick watch doesn't?
Private Sub UpdateCPUUsage(ByVal strSelectedServer As String) 'Performance items come Performance Monitor: perfmon.msc. 'The right-click on columns at bottom and select "add counter" to see list.
Possible Duplicate: Memory leak in .net application I am working on a desktop application in VB.net 2005.The application contains a timer with an interval of 1 min. Each time the timer ticks, a set of functions gets executed, mostly database related. Initially the application runs fine. In processes(Task manager) the cpu usage goes to 100% every time the timer is invoked. But the timespan is around 1 sec(negligible). However as the time passes and after around 20 hours the time span of timer_tick increases to something like 20-30 secs. In this period cpu usage is 100% and the application does not responds.Gradually the time span of timer_tick increases to 1 min and the cpu uses gets stuck to 100% and the application does not responds. All objects are properly disposed. Moreover, this issue is with pentium 4 processors. The application runs fine on core 2 duo.I am using DevEx controls in my application.The program runs fine with less records in database.I have run the CLR Profiler. The code seems to be fine.
I want to get the log entries of "Applications and Services LogsMicrosoftWindowsDiagnostics-PerformanceOperational", do not know how to do that. I know how to get System event log:
Dim myEventLog As New EventLog("System", ".", "EventLog") Dim myLogEntryCollection As EventLogEntryCollection = myEventLog.Entries For Each objELE As EventLogEntry In myLogEntryCollection
i have a large vb.net application, i wanna analyze the performance of the program at runtime.for example, i want to know how many times a certain function was called, or the usage of computer resources for eah function,
I have a dataset in a Visual Studio 2010 Web App project which accesses the DB with a complex SQL statement. If I run the statement in SQL Management Studio directly, it loads in a less than a second. If however, I run it using the "Preview Data" button in the dataset designer, or I try to access it on a page (with a gridview for example), it takes over 40 seconds!
I want to get the log entries of "Applications and Services LogsMicrosoftWindowsDiagnostics-PerformanceOperational", do not know how to do that. I know how to get System event log: Dim myEventLog As New EventLog("System", ".", "EventLog") Dim myLogEntryCollection As EventLogEntryCollection = myEventLog.Entries For Each objELE As EventLogEntry In myLogEntryCollection .......... Next
I want to keep a track which pages are mostly accessed and heavy traffic.And page can be accessible to more user at any instant of time. As number of users increased then login page is not loading and site goes very slow