Search Results

Search found 52432 results on 2098 pages for 'web tools'.

Page 171/2098 | < Previous Page | 167 168 169 170 171 172 173 174 175 176 177 178  | Next Page >

  • Subscribe/Publish Model in Web-based Application (c#) - Best Practices for Event Handlers

    - by KingOfHypocrites
    I was recently exposed to a desktop application that uses an publish/subscribe model to handle commands, events, etc. I can't seem to find any good examples of using this in a web application, so I wonder if I am off base in trying to use this for web based development (on the server side)? I'm using asp.net c#. My main question in regards to the design is: When using a publish/subscribe model, is it better to have generic commands/events that pass no parameters and then have the subscribers look at static context objects that contain the data relevant to the event? Or is it better to create custom arguments for every event that contain data related to the event? The whole concept of a global container seems so convenient but at the same time seems to break encapsulation. Any thoughts or best practices from anyone who has implemented this type of model in a web based application? Even suggestions on this model out of the scope of my question are appreciated.

    Read the article

  • Robust way to keep records of software releases?

    - by japreiss
    We release a number of small plug-ins that go along with our software. Each plug-in allows our software to talk to a single manufactuer's hardware. I would like to devise a system for keeping track of plug-in releases. Example info that should be stored: Hardware manufacturer name 32-bit? 64-bit? both? What modes of operation does the hardware support? What versions of the manufacturer's driver have been tested with the plugin? Desirable properties of the system: Able to synchronize with version control software Stores data in human-readable text file (also good for differ tool) Free visual, spreadsheet-like editor available Able to do simple analysis like "What is the oldest plug-in?" I've got to imagine that someone else has tackled this problem already. Right now my best guess is XML/JSON with a visual editor, but I have been disappointed in the editors I've tried so far. I'd like to get input from some more experienced developers. Thanks!

    Read the article

  • SharePoint 2007: Error message when adding a Web part to a list or library page

    - by Cherie Riesberg
    Sytmptom:   You have a list or document library page (allitems.aspx) and you are trying to add a Web part.  You get an error message: Unable to add selected Web part(s).  (name of Web part): The file is not checked out.  You must first check out this document before making changes. Publishing features are not turned on and it is not a page that lives in a library accessible by the GUI. Solution:  Open the site in SharePoint Designer and check out the page. Then, check in the page after modifications are done.  It seems like this is just a bug.

    Read the article

  • Frequency to submit sitemap to search engines

    - by user577691
    i have went live with my site and being new to search engines and SEO fields not sure what should be the best way to handle sitemap.xml. I have created sitemap.xml and submitted it to Google using webmaster tool Yahoo/Bing using Bing Webmater Ask.com now since site will get updated every 2-3 times per week i am not sure what should be the best approach. Do i need to submit sitemap.xml again If i need to submit sitemap.xml again and again what should be the frequency to submit that Please suggest the best approach

    Read the article

  • Alternatives to OAuth?

    - by sdolgy
    The Web industry is shifting / has shifted towards using OAuth when extending API services to external consumers & developers. There is some elegance in simple....and well, the 3-step OAuth process isn't too bad ... i just find it is the best of a bad bunch of options. Are there alternatives out there that could be better, and more secure? The security reference is derived from the following URLs: http://www.infoq.com/news/2010/09/oauth2-bad-for-web http://hueniverse.com/2010/09/oauth-2-0-without-signatures-is-bad-for-the-web/

    Read the article

  • Website (X)HTML Code Change Detection [closed]

    - by 0pt1m1z3
    I am looking for an enterprise-grade service or a tool that can be used to scan / fingerprint websites and notify when major XHTML code changes are detected. The tool should be able to continuously scan thousands of websites and determine the percentage of HTML code that has been modified since the last run. And then either save the data where it can be easily accessed or send periodic notifications. I know of services like ChangeDetect.com, but they don't do markup only changes and instead focus on everything, including content. We don't really care about presentation content, because a lot of sites we need to cover are updated frequently with content.

    Read the article

  • Is there a tool for detecting sites using the same template?

    - by KTB
    I often buy webtemplates online, and when I do so I often look at the demos to get some inspiration on how to use the components. But I would love to look at other sites which have already implemented a full website using this template. So I am looking for a tool which searches sites having a similar HTML as the demos (and therefore are probably implementing the template). I am referring to templates which do not have a static text like "Created with Template FooBar by BarFoo" in the footer.

    Read the article

  • Is there a way I can verify my Google Analytics custom report?

    - by SnowboardBruin
    I want to track scrolling on my website since it's a long page (rather than multiple pages). I saw several different methods, with and without an underscore for trackEvent, with and without spaces between commas <script> ... ... ... ga('create', 'UA-45440410-1', 'example.com'); ga('send', 'pageview'); _gaq.push([‘_trackEvent’, ‘Consumption’, ‘Article Load’, ‘[URL]’, 100, true]); _gaq.push([‘_trackEvent’, ‘Consumption’, ‘Article Load’, ‘[URL]’, 75, false]); _gaq.push([‘_trackEvent’, ‘Consumption’, ‘Article Load’, ‘[URL]’, 50, false]); _gaq.push([‘_trackEvent’, ‘Consumption’, ‘Article Load’, ‘[URL]’, 25, false]); </script> It takes a day for counts to load with Google Analytics, otherwise I would just tweak and test right now.

    Read the article

  • Is there a better strategy than relying on the compiler to catch errors?

    - by koan
    I've been programming in C and C++ for some time, although I would say I'm far from being an expert. For some time, I've been using various strategies to develop my code such as unit tests, test driven design, code reviews and so on. When I wrote my first programs in BASIC, I typed in long blocks before finding they would not run and they were a nightmare to debug. So I learned to write a small bit and then test it. These days, I often find myself repeatedly writing a small bit of code then using the compiler to find all the mistakes. That's OK if it picks up a typo but when you start adjusting the parameters types etc just to make it compile you can screw up the design. It also seems that the compiler is creeping into the design process when it should only be used for checking syntax. There's a danger here of over reliance on the compiler to make my programs better. Are there better strategies than this? I vaguely remember some time ago an article on a company developing a type of C compiler where an extra header file also specified the prototypes. The idea was that inconsistencies in the API definition would be easier to catch if you had to define it twice in different ways.

    Read the article

  • What are some advantages / disadvantages to working on a remote development machine?

    - by robertpateii
    At home I have a fast rig with my dev environment running in virtual box. that works great, but at work I have a so-so laptop that can barely push visual studio express, outlook, and a dozen chrome windows at the same time. So I can either ask for a dedicated desktop to do development on, or I can ask for a slice on an existing server from IT and remote into it. Setup-wise, the remote option is faster and cheaper. But I don't know its affect on production in the long term. I've done small amounts of work through a remote connection, but never extended development. Do you have experience with this? What are some of the ads/disads to it? Did it make you less productive?

    Read the article

  • Difference in stats on the same domain - seo related

    - by Duoweb
    Recently i started to optimize my company's website. With a little help of market samurai. One of the adjustments was a www redirect to the prefered domain. We currently rank on #8 for our desired keyword with the domain without the www. If I add a custom url (our domain with www), we see a difference in stats BLP (backlinks to the page) {it's the website on top} Can anyone tell me how i could merge this so we might even rank better for our keyword? What I did so far: 3 weeks ago set the www.duo.be as prefered domain in GWT and put a 301 redirect in place so the duo.be gets redirected to www.duo.be Both domains are in GWT and both with www.duo.be as preferred domain. Tho I see no change. Can anyone help me out, because i'm clueless. Marketsamurai screenshot: http://i.stack.imgur.com/CYkb8.png

    Read the article

  • After a domain change, what can I do to recover lost traffic, rankings, impressions etc? [duplicate]

    - by Felix
    This question already has an answer here: How do I rename a domain and preserve PageRank? 3 answers I moved my site to a legacy exact-match domain I purchased about a couple of months ago. I have seen significant reduction in traffic, impressions, and rankings. I did all the right steps/best practices: change of address in GWT, map old site hierarchy and match to new site for 301 redirects etc. Indexation has gone through the Google process: old site has all but dissappeared from he index and new site is indexed, albeit with some 404 errors which I am addressing. Does anyone else who has gone through eh domain change process have any thoughts/advice? Thanks!

    Read the article

  • Is there a version control system that can show changes to a specific method or function?

    - by chesles
    Sometimes it would be nice to be able to say something like: (git|svn|hg|etc) diff Foo.c:main (git|svn|hg|etc) log log Foo.c:main to see the changes made to a specific function within a source file since the last commit, or the complete history of changes. My question is two-fold: Does something exist that does this? Would such a tool be practical? It would have to do some simple parsing of the code at each revision in order to compare different versions of the function; would the overhead be too much for it to be efficient?

    Read the article

  • Azure, a Beautiful Color, and So Much More...

    - by KKline
    Windows and SQL Azure Resources Cloud computing is more than just the latest buzz word in the IT trade papers. It is a remarkable paradigm shift with as much potential to "turn over the apple cart" of IT computing as client-server had for the world of mainframe and minicomputers. If you're not spending time to learn about cloud computing, in general, and SQL Azure, in particular, then you're missing the boat in a big way. (Ha! Two big metaphors in one afternoon. My high-school English teacher would...(read more)

    Read the article

  • Change from static HTML file to meta tag for Google Webmaster verification

    - by Wilfred Springer
    I started verifying the server by putting a couple of static HTMLs in place. Then I noticed that Google wants you to keep these files in place. I didn't want to keep the static HTMLs in, so I want to switch to an alternative verification mechanism, and include the meta tags on the home page. Unfortunately, once your site is verified, you never seem to be able to change to an alternative way of verification. I tried removing the HTML pages. No luck whatsoever. Google still considers the site to be 'verified'. Does anybody know how to undo this? All I want to do is switch to the meta tag based method of site ownership verification.

    Read the article

  • Issue with sitemap in GWT

    - by Anusha
    I have an e-commerce website www.beyondtime.in, i have been constantly monitoring the google bot crawling on my website and my webmaster account. Lately, i have found two issues that i have not been able to understand and hence want your help. 1.) The Google Bots have been only crawling www.beyondtime.in/telecom.php this URL of my website, when the URL is not even valid. So, kindly help me understand what needs to be done to let Google crawl other pages of the website as well. 2.) The second question is about the Google Webmaster account, where i've submitted my sitmap with 227 URLs, but out of that only 156 have been indexed. Also none of the images of my website have been indexed by Google. So kindly help me with this as well. Thanks

    Read the article

  • SQL Live Monitor

    - by TiborKaraszi
    I just found this one out there and wanted to share it. It connects to an instance and show you a bunch of figures. Nothing you can't extract yourself with SQL queries, but sometimes it is just nice to have one tool which is very easy to use. Here's what it looks like when connecting to an instance with no load on it: As you can see, there are some hyperlinked pages as well, and there are also some interesting options (like logging to CSV or for PAL analysis) under the "Option" button. One more thing...(read more)

    Read the article

  • What Are Keyword Tools and How Do You Use Them?

    Keyword Tools are a group of tools that provide analysis and reporting on the use of keywords or phrases on your (or your competitors) website. They are a vital set of tools in terms of optimising your website for high search engine ranking and also for analyzing your competitor's websites.

    Read the article

  • GDD-BR 2010 [2C] Building for a Faster Web

    GDD-BR 2010 [2C] Building for a Faster Web Speakers: Eric Bidelman Track: Chrome and HTML5 Time: C [12:05 - 12:50] Room: 2 Level: 151 Why should a web app be less performant than a native app? This session will focus on creating the next generation of web applications. We'll look HTML5 features that increase app performance, Chrome Developer Tools to speed development, Native Client for running native C++ in the browser, and Google Chrome Frame to bring these awesome features to users with older browsers. From: GoogleDevelopers Views: 84 1 ratings Time: 37:32 More in Science & Technology

    Read the article

  • How to find the reason for a weekly downtime on an Ubuntu web server hosted by AWS?

    - by IceSheep
    We started monitoring our web server using Pingdom and found out that we have a downtime of a few minutes every Sunday at 0:00 UTC. The test runs every minute and checks if a successful HTTP response (code 200) is returned on port 80. The test fails due to a timeout (no response after 30 seconds). Here's what we've already checked – without success: Since we run our webserver behind a load balancer, I've set the Pingdom test on the load balancer's public DNS and the webserver's public DNS in order to find out if there's a problem with the AWS load balancer – both tests return the same result We set up Munin on our webserver. Everything looked fine even after the failure. Since the last failure lasted only 2 minutes I suppose Munin couldn't capture a potential problem (it only checks every 5 minutes) I have checked /var/log/apache2/error.log and /var/log/syslog for suspicious entries I have checked /etc/cron.weekly and /etc/crontab for suspicious entries I have searched for files created or last-modified during 0:00 and 0:15 using this method: touch -t 201209020000 start touch -t 201209020015 end find / -newer start -and ! -newer end (nothing found) Has anybody experienced a similar problem? Any proposals on how to find the reason for this behavior? It's Ubuntu 10.04 LTS running on an AWS m1.large instance. Thanks!

    Read the article

  • How can I get stats for what 3rd-party sites have embedded our iframe widget?

    - by Su'
    Say we've produced a widget for other sites to use, like so: <iframe src="http://example.com/whatever.php" frameBorder="0" width="200px" height="300px" scrolling="no"></iframe> The client would like to be able to see within GA who has embedded the thing. Is there some referer information automatically passed that I can look for, or do I need to add something? whatever.php is already loading the analytics Javascript(we're also tracking clicks on an outbound link). [EDIT] Looking around a bit more, I found what seems to be a similar question on SO with an answer saying this can be found, automatically, but I still can't seem to find the information. The question's also old enough the respondent is probably referring to the old interface, though. Maybe someone could explain getting to it in the new look. (I won't likely be able to train this client to switch, deal with the old look, etc.)

    Read the article

  • Framework to implement an in game gui editor

    - by momboco
    I need to do an in game gui editor. The game engine has his own widgets elements and I don't want a gui library that substitute it. The most difficult task is the implementation of the functionality that makes usable to artists and designers. Positioning Resize Alignment between some elements Multiselection Relationship between children and parents Add guides Magnet to place elements quickly Use of layers Undo / Redo ... I'm searching a framework or something like, with these functionalities implemented. And a form of append my own engine to make use of it. It would be ideally a mixing between a tool like Photoshop and libRocket ( engine rendering independent ).

    Read the article

  • The Connection String Reference Site

    - by Yousef_Jadallah
    In this great site http://www.connectionstrings.com/ you can find about 517 connection strings and 120 providers, drivers and class libraries listed in the database. These for Database Servers as well Data Files. Just you need to choose the needed element then you will get all the information that you calling for.   Hope this helps,

    Read the article

  • Why am I getting this message "Some important page is blocked by robots.txt"?

    - by Rounak
    My site's URL is: www.hackinguniverse.org From some day in Google Webmaster tool a message is showing that says "Some important page is blocked by robots.txt". My robots.txt is: User-agent: Mediapartners-Google Disallow: User-agent: * Disallow: /search Allow: / Sitemap: http://www.hackinguniverse.org/feeds/posts/default?orderby=updated Now for other information - I hosted this website on blogger. I have some other sitemaps too where I had only included "/Search" in Disallow like in this robots.txt file. But those sites are ok. I mean no message is showing on those. So why am I getting that message telling that I have blocked some important page via robot.txt?

    Read the article

  • Modern website/webapp setup

    - by onepiece
    I'm new to web development. From looking at popular open-source frameworks for both front-end and back-end, I have a general idea of what the modern full-stack web setup looks like: Database <- Back-end language ~ REST API <- Front-end Notes: The back-end language (Python, Ruby, PHP, Java) generates the API, which is the only layer between the back and the front. The API will have authentication to protect private data. The front-end sends GET and POST requests to the API. A MVC framework can be used, such as Backbone, Angular, or Ember.js. Does this align with best practices for web development?

    Read the article

< Previous Page | 167 168 169 170 171 172 173 174 175 176 177 178  | Next Page >