WCF / ASMX :: Persistent HTTP - Connection For All Calls To A WS?
Apr 25, 2010
I need to connect to a WebService provided by someone else. This WS (all https) has three methods that are accessible without having logged on:
login, logoff, getVersion . All other methods require that the login-method has been called before. Nothing special unto this point. However, the docs state that I have to make sure that (quote):
"All method calls between login and logoff are to be carried out by means of the same persistent http-connection (key word: http persistent connections or http connection reuse"
I seem to be unable to figure out how that would work - all WS'es I ever utilized were either taking user/pwd combinations in each method call or the login-method returned something like a SessionID which was then used for each subsequent call to a given method (i.e. passed as a parameter). If I call the login-method and subsequently call any method that requires authentication, the call will fail with an exception telling me that I need to log in first.
I want my cookie to disappear when the user closes their brower-- I've already set some promising looking properties, but my cookies pop back to live even after closing the entire browser.
HttpCookie cookie = new HttpCookie("mycookie", "abc"); cookie.HttpOnly = true; //Seems to only affect script access cookie.Secure = true; //Seems to affect only https transport
What property or method call am I missing to achieve an in memory cookie?
I have got ssl security for my website.I have written onLoad that [URL] ... Should redirect on [URL] ... it redirects but first the [URL] ....pages loads full and then after a sec it redirects...I want to solve this
I have a page that calls on an asmx within the project (for some ajax stuff). Within that asmx I'm calculating a value and returning it to be output to the screen.
However I need to do a recalculation that includes summing this value with a few others. The function that performs the calculation is in a UserControl on the page.
How do I get a reference to the page object so I can use FindControl to get an instance of the user control so I can call that calcuation function?
I am working on web application in which all of the data I am submitting back to the server is being done on the client side using JQuery AJAX calls to my .NET web services using JSON. This works perfectly to enhance my user experience and greatly reduces post backs, etc.What I am looking for suggestions on is securing the web services being exposed to the client side from being used by anything but my JQuery code on my site. The security hole that I see here is that since the client-side JQuery can access the web services, so could anything.
I can't seem to call a web service method from Ajax with both POST and GET.
Initially only the POST would work and GET would causes this error:
{"Message":"An attempt was made to call the method u0027getDatau0027 using a GET request, which is not allowed.","StackTrace":" at System.Web.Script.Services.RestHandler.GetRawParams(WebServiceMethodData methodData, HttpContext context)at System.Web.Script.Services.RestHandler.ExecuteWebServiceCall(HttpContext context, WebServiceMethodData methodData)","ExceptionType":"System.InvalidOperationException"}
I fixed that by adding this attribute: [ScriptMethod(UseHttpGet=true)] but now GET causes this error:
[code]....
So is it true that you can only use either POST or GET and not both from Ajax?
I have an XML web service where the authentication used is BASIC WINDOWS. Following is my client code to connect to the service.
[Code]....
Now, the issue is that the client is really authenticated only in the first attempt(First time I execute the code in VS). Further attempts succeed even if the code to assign credentials is commented or even after providing wrong user id/ password. I ensured that the server is hit each time(Just to make it clear that it not just because of client side caching).Is this the behavior of Basic authentication/credential cache. I would prefer the client be authenticated each time he hits the server.
Our site was under Slow HTTP POST DOS attack recently. It's when your server receives a lot of incoming connections and keep them for long time because the sender sends information very slow and server cant serve real users' request - denial of service.
The problem is that I can't drop the HTTP connection in asp.net:
Response.End(), throw new Exception(), even Thread.CurrentThread.Abort()
don't close the connection. It waits while sender sends all the fake data, then answer to it with '500 server error' or something. The slow POST attack still will be successful in this case.
I need to know the best way to do the following. I have nested business level APIs (say level 1 & level 2). L1 needs to call L2. Both APIs use the database layer directly at their own nesting levels.
Now, in the database layer, I fetch the db connection from the pool each time as follows:
SqlConnection conn = new SqlConnection(connString); conn.Open();
Is it proper to fetch the db connection each time on every DB level call as above? I know it will return a connection from the ASP.NET connection pool. However, wouldn't it be better to maintain the same DB connection throughout the nested calls (or throughout the current http request lifetime)? Will fetching a connection from the pool each time cause issues with nested TransactionScopes?
I'm looking into building an ASP.NET MVC application that exposes (other than the usual HTML pages) JSON and XML REST services, as well as Web Sockets. In a perfect world, I would be able to use the same URLs for the Web Sockets interface as I do for the other services (and determine which data to return by what the user agent requests) but, knowing that IIS wasn't built for persistent connections, I need to know if there's a way that I can accept (and possibly even handshake) the Web Sockets connection and then pass the connection off to another service running on the server. I do have a workaround in mind if this isn't possible that basically involves using ASP.NET to check for the Web Sockets connection upgrade headers, and responding with a HTTP/1.1 302 Found that points to a different host that has my Web Sockets service configured to directly listen to the appopriate endpoint(s).
When using .net remoting, does the server limit incoming client remoting calls?
I find a particular remoting call (during ASP.NET page rendering) to take from 200ms to 1500ms. While the underlying data call is only 50ms. Factoring in remoting overhead of 150ms per call, the only difference between the two cases is that the latter scenario has about a dozen more parallel remoting calls in progress. So my guess is that when too many remoting calls are happening, some will get queued up? I also doubt system resource is the cause of the delay because it is not nearly saturated.
Searching MSDN, I find the below:
http://msdn.microsoft.com/en-us/library/ms973907.aspx clientConnectionLimit: specifies how many connections can be simultaneously opened to a given server. The default is 2. This is exactly the same as the connection limit on ServicePoint in the net classes.
That seems awfully low to me and if were the case, my app's performance would have been much worse. Can someone confirm if there is indeed a connection limit or some other throttling in .net remoting?
On an ASP.NET Web Application I have an upload file functionality. I'm restricting the file size up to 10 MB by configuring the inside web.config the maxAllowedContentLength attribute. (I'm using IIs 7.0 BTW).
It get the desired HTTP Error when I access the application from the local machine:
HTTP Error 404.13 - Not Found The request filtering module is configured to deny a request that exceeds the request content length.
When I access the application from another machine I get
The connection was reset The connection to the server was reset while the page was loading.
How can I get a more descriptive error when my application is accessed remotely?
I made a site www.hediyelikdukkani.com and I added the site to goole from web masters tools. But google doesn't index my web site. When I test my site how the google see from Test like a Google bot from web masters tools it says that
This error stopping my work in my crucial time. The error is "The underlying connection was closed: The connection was closed unexpectedly.".There are multiple WCF service I created in my project. But when I am requesting only one service this error is coming. I compare with other service, no difference. The service works fine in the morning, but afternoon it stops working and raising error.
I installed Silverlight 3.0 Tools for 2008 Sp1, Silverlight 3.0 Developer, WCF RIA Beta and Silverlight 4.0 runtime in my system in the morning. After getting this error I uninstalled WCF RIA Beta. But same issue.Could not find any solution. Please assist me how to solve this issue.
I have a load balancer in front of my web server performing ssl offloading.
so e.g https://my.website.com/service.svc will hit my server's http://my.website.com/service.svc
Now the problem is, http is not available for this url. so when i tried to add service reference from visual studio, i enter the https url, the output from the page refers to http links which does not exist and cause the service reference to fail.
i've already added the neccessary settings inside web.config to make the wcf service work with ssl offloading
<security mode="None"> <transport clientCredentialType="None" proxyCredentialType="None" realm="" /> <message clientCredentialType="UserName" algorithmSuite="Default" /> </security>i've also updated the svc.cs file to [ServiceBehavior(AddressFilterMode = AddressFilterMode.Any)]so my question now is.. how do i make this url return https urls rather than the http ones
The page in which this Page_Load resides is called by a parent page. Each time this Page_Load gets called the value of "file" in the query string will be different. So each time I will be receiving the XML contents from a different file which are inserted into the XmlDataSource and then bound against by the Repeater. Getting the XML data from the file works great but when I bind against the repeater that only works the first time through Page_Load. Each time after that when this code is executed the same results from the first XML file are displayed in the repeater.
What am I missing here. How can I get the XML data to be bound against the repeater on each page load instead of just the first one?
I have a webservice (ASP.NET) deployed on a webfarm. A client application consumes it on daily basis. The problem is that some of its calls are duplicated (with difference of milliseconds). For example I have a function Foo(string a,string b). The client app calls this webmethod as Foo('test1','test2') once but my log shows that it is being called twice or sometimes 3 or 4 times randomly. Is this anything wrong with the webfarm or the code? Note that the webmethod has simple straighfarward insert and update statements.
have a web service that I'm trying to consume from a console app through http post. I receive a 500 exception error from the xmlstring parm being passed to the web service. Add the following to web.config of the web service. Although it is not working in debug in vs 2008 as well.
[Code]....
[Code]....
[Code]....
the host file is setup for the web service as well.
My web service uses .Net Framework 2.0. An e-learning application reads and updates an Access database in the web service via Http Post commands. I test the e-learning application and the web service using the version of IIS 7 that came with Windows Vista. I run IIS using localhost in my browser.
All works well. But a friend is having trouble running the same test with my files on her IIS 6 server. So I'm developing a checklist of issues in setting up an IIS 6 server to run a web service. I plan to include all possible causes of problems including permissions, accounts that are needed, etc. Are there resources to which I can go for this checklist?
I'm new Developer in Web Servicei use VS.2005 C# 2.i have 3 Client PC With Webservice Doing some calculation & Server with ASPX pages the server side will call the client LocalHost webService to do some calculation & return Valuethe problem i get error Massege (The request failed with HTTP status 404: Not Found.)
using System.Web; using System.Web.Security; using System.Web.UI; using System.Web.UI.WebControls; using System.Web.UI.WebControls.WebParts; using System.Web.UI.HtmlControls; [code]...
we need to expose a web service to the 3rd party vendor which can only call using POST protocol, NOT SOAP.So my question, is it better to build a web service or just simply .aspx page?Also is there any best practice documentation on this.
I am after a tool that allows me to catch HTTP traffic on my local machine that is generated through web service calls. I have tried Fiddler, but it doesn't seem to catch my web services that are sent via...
[Code]....
Maybe fiddler can be set up to allow this, I am not sure :-s If anyone knows, can you explain, or if anyone has used a tool on their local machine that can catch HTTP request/responses from the above code?