These days I started new ground in learning intellectual website for example I want to understand the number and name and type of tables and columns and make button and interface of website with all aspx and cs and classes and to cut long story short blank solution and sth like this but i do not know where I should start .I prefer to have suggestion of some good books whith functional example I searched in alot of website but I could not find anything ,now i am about losing my time and energy
Iīm trying to automate a confirmation email going out to a newly registered user.
I've added reference to the interop.outlook library and I import it to the asp webpage. It actually works, so thatīs fine.However I donīt really know whatīs going on here, so maybe you guys can tell me?
1. What needs to be installed on the server for this to work?
2. Who will be the sender? When I do this locally, Iīm the sender, but on the site??
I notice that some sites are coping the content of one of my client's sites using automated agents. I want to detect their requests and show them a captcha code to prevent them from coping the site content. Is there anyway to detect them?
I'm working on a public site that soon will not longer be 100% public. Paid users will be able to access 100% of the features while "anonymous" users will access a reduced version. The problem is that anonymous users and paid users will access the same pages, but they'll see different things. I wonder about how to keep using cache for all pages. I have to 100% sure that I show the correct content for each type of user. Will be extremely bad to show a cached version of the paid site to an anonymous user (and the other way around). Preferably, without adding anything to the url to differentiate user type.
I haven't done anything with mobile devices and was just asked to create a mobile version of one of our sites. Is there anything specific to .net for this? I'm not sure where to start.
I'm working with VS2010 and I wonder should I use one Solution for multiple web applications (web sites) or is it better to create separate web applications (with it's own solution) for each web site?
I am about to a client develop a mobile version of their EPiServer site. The mobile version should have different functionality than the regular but some content might be shared so that it won't have be maintained twice.
I'm thinking about using jQuery Mobile as the user interface system to reach as many users as possible with the same solution and possibly 51degrees.mobi for best detection of mobile devices.
Does anyone have any experience in how to do this with EPiServer? How to structure? I'm thinking of using a sub-node (/mobile) that gets hidden on the regular site but I'm not sure it's the best solution.
After seeing the Hanselman "You are doing it wrong" video I start to use the Web Publish feature of VS2010.
What I'm really missing is that the websites sometimes gives errors while the site is publishing because the feature does not copy the app_offline.htm file to the server.
I do not want to start using MSDeploy scripts, because I have several sites and want to keep it simple.
Maybe there is a simple tweak to tell the wizard to copy and then delete the file.
The website compiled fine before as a regular website in VS 2008. However, in my attempt to convert it into a solution and separate out the business logic into its own project, I'm getting issues where it says that my codebehind references to controls cannot be found in the context. This is for every aspx or ascx file. All my references are there. Is there a correct way to convert a website into a solution? I simply drag-and-dropped the web files into a project in the solution essentially.
I'm trying to create a .NET solution and add class libraries to it. First - in VS2008 I go to File > New > Project > Other Project Types > Visual Studio Solutions > Blank Solution. Then - after VS2008 creates the blank solution I right click on the solution and select Add > New Project > Visual C# > Class Library. However, when I add the Class Library to the solution, the solution icon disappears from Solution Explorer (although the solution name is still in the VS2008 title bar). Why does the solution icon disappear from the Solution Explorer when I attempt to do this?
User registers for the siteAn email is sent to the user confirming their registration am using : Asp.net 3.5(C#)itefinity as a CMSMS SQL 2008 server r2My question is what is the best way to automatically send the email?Should it be done in ASP.net or as a database trigger using something like xp_sendmail
Here's an interesting problem. I have an ETL script written in c# that I have been running manually on a somewhat regular basis. It is used to update my web app's database.
I want to automate the ETL process AND create an interface for the web app admins to manually start the ETL process.
I could have sql server kick off the ETL process on a schedule and implement a button or something on a web page that will do the same thing, but I don't want to put my code in 2 different places because I don't want to update it in 2 places when it changes. But I don't know how to make my web app tell SQL server to manually start a scheduled process. Can this be done? OR
I could somehow implement the scheduling in the web app itself, but by now most people are familiar with the problems that are faced when trying that (app may not be running at certain times, must wait on request to start a process (without some trickery)). Also, since the ETL process takes a while, I don't want to make some poor end user wait on a response, so it would definitely have to use a new thread.
I have a bunch of C# functions with string, int and bool arguments that serve as data entry interfaces. I'd like to be able to create a webform for each method; a textbox for each string/int and a checkbox for each bool.
I am not able to send mails thru my ASP.net pages. If I host site with 172.0.0.1 IP on other hosting server then it will work but if I set the 202.71.148.84 IP provided by my Hosting Provider that time I can't if any other information is required to address this issue.
I'm working on a site that will offer a free service. It's a very simple site: A visitor will choose some text files from their PC, and once he has finished choosing them, he will click on the "Submit" button. This "Submit" button will simply upload the text files to X folder on the server and run a 3rd-party executable that will process these files. That's it. The exe will be doing all the work and creating a new file that the user will be downloading. This is the problem: I've found out the hard way that shared hosting plans don't allow executables on their servers. So it's been suggested that I purchase a VPS or dedicated server plan for this, but the cheapest one I found is $39/mo.
Although my site will have a "premium" service, it's mostly free, and I wasn't expecting to spend too much on hosting. As a matter of fact, I was planning on using my current shared plan for this. My question is: is there any other way of being able to do this? It would need to be automated; being free, tons of people (hopefully) will go to the site, start uploading text files, and wait for the result. I could do it manually (download the text files from the server, run the EXE locally, and finally upload the result and email the user), but that's almost impossible.
I have a class with many properties - strings, booleans, etc. This class was created from Entity Framework.
Now I want to quickly create the controls for the properties. I dont care about linking the controls to an object - I can do this later. I just find it a pain to create 50+ controls on the webpage when the class could be used to create a 'template'.
What are some good automated tools for load testing (stress testing) web applications, that do not use record and replay of HTTP network packets?I am aware that there are numerous load testing tools on the market that record and replay HTTP network packets. But these are unsuitable for my purpose, because of this:he HTTP packet format changes very often in our application (e.g. whenwe optimize an AJAX call). We do not want to adapt all test scripts just becausethere is a slight change in HTTP packet format
I'm developing a system that should send emails to clients on specified dates. i don't have much idea about it.i have created a simple windows service .how to get the email address from the sql server database and attach to windows service?
I have a website which needs to send automated emails. I have the emailing set up fine but what I haven't managed to do is embed an image in the messages. I have tried the following code :
[Code]....
All this does is is place the standard "box with red cross" in the email body and attach a '.dat' file.
I have a problem with using menu control in ASP.NET. I am using a maser page with web.sitemap and menu control. I am designed the whole thing with css but it not looks what I want. I have noticed when I build the page it generates some css in the head section and that srcrew my whole design. How can I reset this automated injected css for my menu to I could use pure CSS for my project.
The data that we display in gridview and details view can easily be scraped using automated tools. My question is if someone uses some automated tool and logs into the website and scraps data how can we find that? How can we detect whether a human is viewing the site or a tool? For example one way is by calculating the time up to which a user stays in page from which we can detect whether human intervention is involved. I do not know how to implement that but just thinking about this method. how to detect and prevent automated tools from scraping data from my website? I have used security image in login section, the user has to log in by entering username, password and security image displayed in the login section. Even then in home page a human may type security image and log in to the site and then use an automated tool to scrap data from the site!!When the recaptcha image appears after a period of time then that time alone the user may type the security image and again use automated tool to scrap data from the site. Actually, I have even developed a tool to scrap data from another site. So only i want to prevent this from happening in my site!