Search Results

Search found 21769 results on 871 pages for 'check constraints'.

Page 420/871 | < Previous Page | 416 417 418 419 420 421 422 423 424 425 426 427  | Next Page >

  • Migrating 2008 AD to Windows SBS 2011 and Forefront TMG 2010

    - by Tong Wang
    Our company has two physical servers: a Dell R710 with 4 NIC and a Dell R410 with 2 NIC. Right now, we have Windows server 2008 R2 installed on R710 with AD setup to host our domain. I am thinking to install both SBS 2011 and TMG 2010 on the other R410 running on Hyper-V. As I am fairly new to Windows Server technology, I'd like to check with you fellow Windows admins and see what is the proper way to install SBS 2011 and TMG 2010 and how to migrate the 2008 AD. Your advice and help will be highly appreciated.

    Read the article

  • The Earth at Night [Video]

    - by Jason Fitzpatrick
    This fresh video from NASA provides the clearest view of the Earth at night ever seen, thanks to the Suomi National Polar-orbiting Partnership Satellite. Check out the video and accompanying pics to see the stunning views. In daylight our big blue marble is all land, oceans and clouds. But the night – is electric. This view of Earth at night is a cloud-free view from space as acquired by the Suomi National Polar-orbiting Partnership Satellite (Suomi NPP). A joint program by NASA and NOAA, Suomi NPP captured this nighttime image by the satellite’s Visible Infrared Imaging Radiometer Suite (VIIRS). The day-night band on VIIRS detects light in a range of wavelengths from green to near infrared and uses filtering techniques to observe signals such as city lights, gas flares, and wildfires. This new image is a composite of data acquired over nine days in April and thirteen days in October 2012. It took 312 satellite orbits and 2.5 terabytes of data to get a clear shot of every parcel of land surface. This video uses the Earth at night view created by NASA’s Earth Observatory with data processed by NOAA’s National Geophysical Data Center and combined with a version of the Earth Observatory’s Blue Marble: Next Generation. Hit up the link below for the full NASA press release, including more videos and photos. How to Factory Reset Your Android Phone or Tablet When It Won’t Boot Our Geek Trivia App for Windows 8 is Now Available Everywhere How To Boot Your Android Phone or Tablet Into Safe Mode

    Read the article

  • Simple way of converting server side objects into client side using JSON serialization for asp.net websites

    - by anil.kasalanati
     Introduction:- With the growth of Web2.0 and the need for faster user experience the spotlight has shifted onto javascript based applications built using REST pattern or asp.net AJAX Pagerequest manager. And when we are working with javascript wouldn’t it be much better if we could create objects in an OOAD way and easily push it to the client side.  Following are the reasons why you would push the server side objects onto client side -          Easy availability of the complex object. -          Use C# compiler and rick intellisense to create and maintain the objects but use them in the javascript. You could run code analysis etc. -          Reduce the number of calls we make to the server side by loading data on the pageload.   I would like to explain about the 3rd point because that proved to be highly beneficial to me when I was fixing the performance issues of a major website. There could be a scenario where in you be making multiple AJAX based webrequestmanager calls in order to get the same response in a single page. This happens in the case of widget based framework when all the widgets are independent but they need some common information available in the framework to load the data. So instead of making n multiple calls we could load the data needed during pageload. The above picture shows the scenario where in all the widgets need the common information and then call GetData webservice on the server side. Ofcourse the result can be cached on the client side but a better solution would be to avoid the call completely.  In order to do that we need to JSONSerialize the content and send it in the DOM.                                                                                                                                                                                                                                                                                                                                                                                            Example:- I have developed a simple application to demonstrate the idea and I would explaining that in detail here. The class called SimpleClass would be sent as serialized JSON to the client side .   And this inherits from the base class which has the implementation for the GetJSONString method. You can create a single base class and all the object which need to be pushed to the client side can inherit from that class. The important thing to note is that the class should be annotated with DataContract attribute and the methods should have the Data Member attribute. This is needed by the .Net DataContractSerializer and this follows the opt-in mode so if you want to send an attribute to the client side then you need to annotate the DataMember attribute. So if I didn’t want to send the Result I would simple remove the DataMember attribute. This is default WCF/.Net 3.5 stuff but it provides the flexibility of have a fullfledged object on the server side but sending a smaller object to the client side. Sometimes you may hide some values due to security constraints. And thing you will notice is that I have marked the class as Serializable so that it can be stored in the Session and used in webfarm deployment scenarios. Following is the implementation of the base class –  This implements the default DataContractJsonSerializer and for more information or customization refer to following blogs – http://softcero.blogspot.com/2010/03/optimizing-net-json-serializing-and-ii.html http://weblogs.asp.net/gunnarpeipman/archive/2010/12/28/asp-net-serializing-and-deserializing-json-objects.aspx The next part is pretty simple, I just need to inject this object into the aspx page.   And in the aspx markup I have the following line – <script type="text/javascript"> var data =(<%=SimpleClassJSON  %>);   alert(data.ResultText); </script>   This will output the content as JSON into the variable data and this can be any element in the DOM. And you can verify the element by checking data in the Firebug console.    Design Consideration – If you have a lot of javascripts then you need to think about using Script # and you can write javascript in C#. Refer to Nikhil’s blog – http://projects.nikhilk.net/ScriptSharp Ensure that you are taking security into consideration while exposing server side objects on to client side. I have seen application exposing passwords, secret key so it is not a good practice.   The application can be tested using the following url – http://techconsulting.vpscustomer.com/Samples/JsonTest.aspx The source code is available at http://techconsulting.vpscustomer.com/Source/HistoryTest.zip

    Read the article

  • Der einfache Weg zur Oracle Fusion Middleware

    - by A&C Redaktion
    Was passiert eigentlich mit Produkten und Anwendungen, die unter Oracle Application Server entwickelt und betrieben wurden – jetzt, da der WebLogic Server diesen als wichtigste strategische Middleware-Plattform ablöst? Der Oracle Platinum Partner OPITZ CONSULTING widmet dieser Frage, die in den nächsten Jahren so manches Unternehmen beschäftigen wird, bereits jetzt eine umfassenden Kampagne aus Informations- und Serviceangeboten. Für OPITZ CONSULTING, wie auch andere Partner, die im Bereich Oracle Fusion Middleware tätig sind, gilt es jetzt, betroffenen Kunden die Möglichkeiten und Szenarien eines Upgrades vorzustellen, die Vorteile der neuen Lösungen zu erläutern und gemeinsam abzuwägen, welche Schritte nötig sind, um ihr Unternehmen fit für die Zukunft zu machen. Letztlich zählt die Leistung immer noch als das beste Argument. OPITZ CONSULTING hat jetzt die entscheidenden Informationen und Argumente zum Fusion Middleware Upgrade in einem Fact Sheet zusammengefasst. Von der Analyse über die Bewertung und Empfehlung bis hin zum validen Konzept werden die einzelnen Schritte anschaulich dargestellt. Desweiteren hilft die IT-Beratung ihren Kunden mittels Migrations-Check dabei, Varianten, Kosten und Nutzen der Umstellung auf die neue Technologie abzuwägen. Je nachdem, welche Produkte ein Kunde verwendet, bietet OPITZ CONSULTING unterschiedliche Upgrade-Szenarien an. Hier geht's zum Fact Sheet: „Oracle Fusion Middleware – Upgrade!“

    Read the article

  • Der einfache Weg zur Oracle Fusion Middleware

    - by A&C Redaktion
    Was passiert eigentlich mit Produkten und Anwendungen, die unter Oracle Application Server entwickelt und betrieben wurden – jetzt, da der WebLogic Server diesen als wichtigste strategische Middleware-Plattform ablöst? Der Oracle Platinum Partner OPITZ CONSULTING widmet dieser Frage, die in den nächsten Jahren so manches Unternehmen beschäftigen wird, bereits jetzt eine umfassenden Kampagne aus Informations- und Serviceangeboten. Für OPITZ CONSULTING, wie auch andere Partner, die im Bereich Oracle Fusion Middleware tätig sind, gilt es jetzt, betroffenen Kunden die Möglichkeiten und Szenarien eines Upgrades vorzustellen, die Vorteile der neuen Lösungen zu erläutern und gemeinsam abzuwägen, welche Schritte nötig sind, um ihr Unternehmen fit für die Zukunft zu machen. Letztlich zählt die Leistung immer noch als das beste Argument. OPITZ CONSULTING hat jetzt die entscheidenden Informationen und Argumente zum Fusion Middleware Upgrade in einem Fact Sheet zusammengefasst. Von der Analyse über die Bewertung und Empfehlung bis hin zum validen Konzept werden die einzelnen Schritte anschaulich dargestellt. Desweiteren hilft die IT-Beratung ihren Kunden mittels Migrations-Check dabei, Varianten, Kosten und Nutzen der Umstellung auf die neue Technologie abzuwägen. Je nachdem, welche Produkte ein Kunde verwendet, bietet OPITZ CONSULTING unterschiedliche Upgrade-Szenarien an. Hier geht's zum Fact Sheet: „Oracle Fusion Middleware – Upgrade!“

    Read the article

  • .htaccess - permissions forbidden

    - by user1732521
    I have an error with a new virtual host that I can't figure out.. My .htaccess doesn't have web access (403). [Thu Oct 31 17:51:01 2013] [crit] [client ] (13)Permission denied: /srv/data_disk/www /site.dev/.htaccess pcfg_openfile: unable to check htaccess file, ensure it is readable I have set the permissions for the complete htdocs folder to 755, and to owned by my regular user and group (www-data). I have other vhosts set up with the same user and lesser permissions (rw-rw---) on the .htaccess. The virtual hosts are also setup in the same way.. as far a I can tell.. Thanks!

    Read the article

  • Getting Started with Puppet on Oracle Solaris 11

    - by Glynn Foster
    One of the exciting enhancements with Oracle Solaris 11.2 has been the introduction of Puppet. While upstream Puppet did have some rudimentary support for Oracle Solaris 11, Drew Fisher and Ginnie Wray worked tirelessly to add enhance the Oracle Solaris Puppet offering. We've talked to customers over the past few years and asked them what their problems were and what technologies they were using, particularly for configuration management. Puppet came up time and time again, and it made a huge amount of sense bringing it as a 1st class citizen in the Oracle Solaris platform. So what is Puppet, and why is it useful? To quote from PuppetLabs, the guys who are responsible for creating Puppet: Puppet is a declarative, model-based approach to IT automation, helping you manage infrastructure throughout its lifecycle, from provisioning and configuration to orchestration and reporting. Using Puppet, you can easily automate repetitive tasks, quickly deploy critical applications, and proactively manage change, scaling from 10s of servers to 1000s, on-premise or in the cloud. What's more, with Puppet support for Oracle Solaris, administrators can now manage a completely heterogeneous data center from a single or series of Puppet masters. Better still, it's an excellent tool when combined with our new compliance framework to ensure you're meeting your compliance regulations. We're not stopping there of course, and we'll enhance our offerings over time, and work with PuppetLabs to get some of this support upstream (or into the Puppet Forge). So if you've heard some of the buzz around Puppet and never quite got started, and have some Oracle Solaris real estate that you'd love to manage, check out the Getting Started with Puppet on Oracle Solaris 11 guide.

    Read the article

  • Full-text indexing? You must read this

    - by Kyle Hatlestad
    For those of you who may have missed it, Peter Flies, Principal Technical Support Engineer for WebCenter Content, gave an excellent webcast on database searching and indexing in WebCenter Content.  It's available for replay along with a download of the slidedeck.  Look for the one titled 'WebCenter Content: Database Searching and Indexing'. One of the items he led with...and concluded with...was a recommendation on optimizing your search collection if you are using full-text searching with the Oracle database.  This can greatly improve your search performance.  And this would apply to both Oracle Text Search and DATABASE.FULLTEXT search methods.  Peter describes how a collection can become fragmented over time as content is added, updated, and deleted.  Just like you should defragment your hard drive from time to time to get your files placed on the disk in the most optimal way, you should do the same for the search collection. And optimizing the collection is just a simple procedure call that can be scheduled to be run automatically.   beginctx_ddl.optimize_index('FT_IDCTEXT1','FULL', parallel_degree =>'1');end; When I checked my own test instance, I found my collection had a row fragmentation of about 80% After running the optimization procedure, it went down to 0% The knowledgebase article On Index Fragmentation and Optimization When Using OracleTextSearch or DATABASE.FULLTEXT [ID 1087777.1] goes into detail on how to check your current index fragmentation, how to run the procedure, and then how to schedule the procedure to run automatically.  While the article mentions scheduling the job weekly, Peter says he now is recommending this be run daily, especially on more active systems. And just as a reminder, be sure to involve your DBA with your WebCenter Content implementation as you go to production and over time.  We recently had a customer complain of slow performance of the application when it was discovered the database was starving for memory.  So it's always helpful to keep a watchful eye on your database.

    Read the article

  • ReadyNAS Duo - FTP home network connectivity issue

    - by dwabyick
    I have a ReadyNas Duo that I recently purchased and configured on my local home network. I'm an AT&T DSL subscriber using their default 2701HG-B Gateway (2Wire) wireless router. Basically, I have managed to configure my ReadyNAS, and can access HTTP shares and WINS shares just fine, but cannot get access to it via FTP. I've enabled it in the 'Standard File Protocols' section in the UI, and made a share have read/write default access via FTP. My settings: Port: 21 Authentication mode: User Allow upload resumes: Yes Passive ports: 1024-65535 Masquerade as: (empty, but I've tried the internal IP address). I've even poked around at my router, which has a firewall, and even tried to poke a hole for the 'FTP Server'app for the ReadyNAS. Doesn't seem like I should need to do this, or should. Any ideas what else to check? Thanks! -Daniel

    Read the article

  • mount network drive

    - by CaptnLenz
    since i updated my ubuntu to natty narwhal(from 10.04), my mount script doesn't work anymore. The scripts mounts a folder from a NAS (WD mybookworld) in the local network to a folder in my home folder. script looked like that: #!/bin/bash sudo mount //192.168.2.222/Public/Shared\ Music/ /home/simon/Musik/ error: mount: wrong fs type, bad option, bad superblock on //192.168.2.222/Public/Shared Music/, missing codepage or helper program, or other error (for several filesystems (e.g. nfs, cifs) you might need a /sbin/mount.<type> helper program) Manchmal liefert das Syslog wertvolle Informationen – versuchen Sie dmesg | tail oder so now, because the script doesn't work anymore i decided to add the mount-process to my fstab, because the network drive should be mounted on every startup. My fstab entry looks like this: //192.168.2.222/Public/Shared\ Music/ /home/simon/Musik cifs credentials=/home/simon/.smbcredentials 0 0 But it doesn't work, too. I get a message during the startup process, that Musik couldn't be mounted. Are there any log files i can check for errors? The system is a fresh installed 11.04. Greetings

    Read the article

  • How should I design a correct OO design in case of a Business-logic wide operation

    - by Mithir
    EDIT: Maybe I should ask the question in a different way. in light of ammoQ's comment, I realize that I've done something like suggested which is kind of a fix and it is fine by me. But I still want to learn for the future, so that if I develop new code for operations similar to this, I can design it correctly from the start. So, if I got the following characteristics: The relevant input is composed from data which is connected to several different business objects All the input data is validated and cross-checked Attempts are made in order to insert the data to the DB All this is just a single operation from Business side prospective, meaning all of the cross checking and validations are just side effects. I can't think of any other way but some sort of Operator/Coordinator kind of Object which activates the entire procedure, but then I fall into a Functional-Decomposition kind of code. so is there a better way in doing this? Original Question In our system we have many complex operations which involve many validations and DB activities. One of the main Business functionality could have been designed better. In short, there were no separation of layers, and the code would only work from the scenario in which it was first designed at, and now there were more scenarios (like requests from an API or from other devices) So I had to redesign. I found myself moving all the DB code to objects which acts like Business to DB objects, and I've put all the business logic in an Operator kind of a class, which I've implemented like this: First, I created an object which will hold all the information needed for the operation let's call it InformationObject. Then I created an OperatorObject which will take the InformationObject as a parameter and act on it. The OperatorObject should activate different objects and validate or check for existence or any scenario in which the business logic is compromised and then make the operation according to the information on the InformationObject. So my question is - Is this kind of implementation correct? PS, this Operator only works on a single Business-wise Operation.

    Read the article

  • Configure USB GSM Modem

    - by adisembiring
    I want to configure EVDO Usb Modem in Ubuntu 10.10 I insert my usb modem to laptop and check the usb is detected or not using $sudo lsusb and the result is: Bus 003 Device 001: ID 1d6b:0001 Linux Foundation 1.1 root hub Bus 005 Device 001: ID 1d6b:0001 Linux Foundation 1.1 root hub Bus 002 Device 002: ID 201e:2009 Bus 002 Device 001: ID 1d6b:0001 Linux Foundation 1.1 root hub Bus 004 Device 001: ID 1d6b:0001 Linux Foundation 1.1 root hub Bus 001 Device 003: ID 14cd:6600 Super Top USB 2.0 IDE DEVICE Bus 002 Device 002: ID 201e:2009 is my us device. and than I execute command $dmesg | grep -e tty [ 0.000000] console [tty0] enabled [ 818.054660] usb 2-1.3: GSM modem (1-port) converter now attached to ttyUSB0 [ 818.055125] usb 2-1.3: GSM modem (1-port) converter now attached to ttyUSB1 [ 818.055647] usb 2-1.3: GSM modem (1-port) converter now attached to ttyUSB2 [ 818.330641] option1 ttyUSB0: GSM modem (1-port) converter now disconnected from ttyUSB0 [ 818.330743] option1 ttyUSB1: GSM modem (1-port) converter now disconnected from ttyUSB1 [ 818.330840] option1 ttyUSB2: GSM modem (1-port) converter now disconnected from ttyUSB2 [ 1054.917473] usb 2-1.2: GSM modem (1-port) converter now attached to ttyUSB0 [ 1054.917995] usb 2-1.2: GSM modem (1-port) converter now attached to ttyUSB1 [ 1054.918481] usb 2-1.2: GSM modem (1-port) converter now attached to ttyUSB2 [ 1055.214087] option1 ttyUSB0: GSM modem (1-port) converter now disconnected from ttyUSB0 [ 1055.214221] option1 ttyUSB1: GSM modem (1-port) converter now disconnected from ttyUSB1 [ 1055.214356] option1 ttyUSB2: GSM modem (1-port) converter now disconnected from ttyUSB2 Why converter disconnected from ttyUSB2 ? and than I try to execute command $sudo wvdialconf sorry, no modem was detected! Is it in use by another program ? did you configure it properly with setserials ? I change the same type usb modem from my friend, but I still get error above.

    Read the article

  • Troubleshooting Amazon EC2 reboot

    - by tgm
    We've had a server (CentOS) running in EC2 for a few months. It had been going pretty smoothly until today when we got an alarm that the server was unavailable (HTTP service couldn't be reached). So I tried SSHing into the box but that timed out as well. I logged into the EC2 console and it said the instance was running and there wasn't anything in the system log. One odd thing I noticed is that even though we have an Elastic IP attached to it (which shows in the Elastic IP management area), the instance detail is not showing that there is an EIP associated with the instance. I looked through the message log and the last thing I see around the time we got our alert was the dhclient renewed the lease. I'm guessing there may have been some sort of issue with the networking. How might I check if that was the problem, or if there were any other issues that may have caused our instance to stop responding?

    Read the article

  • Dual monitors though 1 HDMI port

    - by Carlos
    I currently have a Dell Studio XPS 13 laptop connected to a 24" HP monitor (w2448hc). Im thinking on getting a second one, however am wondering what i need for the setup (hardware wise). Also I was wondering it there is any down side to it, or something i should be aware of. For example, image quality loss, GPU overloading, or anything important I should know. More than anything Im interested in your advice. Also the monitors do have built in speakers (HDMI sound output), is the sound going to be reproduced by only one monitor or both? Specs Model: Dell Studio XPS 13 OS: Genuine Windows® 7 Home Premium 64-Bit CPU: Intel® CoreTM 2 Duo P8600 (2.4GHz, 3MB L2 Cache, 1067MHz FSB) Chipset: NVIDIA® GeForce® MCP79MX RAM: 4GB 1067MHz DDR3 SDRAM Graphics: SLi NVIDIA® GeForce® 9500M - 256MB Thanks for your advice, if there is anything additional i need to buy an you have a personal preference pass the brand name so i can check it out. Thanks!

    Read the article

  • lsass.exe memory leak on windows 2003 server

    - by thelsdj
    In the past month or so I noticed that lsass.exe has started to leak memory, getting to 500MB+ of ram in under a week after reboot. Before this I had never noticed it using any significant amount of memory compared to other processes on the system. This is happening on 2 identical servers, neither of which has anything to do with Active Directory. Maybe a recent Windows Update has caused this? Any thoughts on things to check? As a side question is there some way to recycle the memory usage of lsass.exe without rebooting? Edit: Here is what I'm seeing in Process Monitor, there are thousands of registry open/query/close a minute from lsass.exe. How can I track down what is triggering these?

    Read the article

  • Error installing php extension OAuth via pecl

    - by PJ
    I'm trying to install the php extension OAuth in my local environment. php.net suggests it's super easy. You just run pecl install oauth. I tried this, and here is the output in terminal: downloading oauth-1.0.0.tgz ... Starting to download oauth-1.0.0.tgz (42,834 bytes) ............done: 42,834 bytes 6 source files, building running: phpize grep: /usr/include/php/main/php.h: No such file or directory grep: /usr/include/php/Zend/zend_modules.h: No such file or directory grep: /usr/include/php/Zend/zend_extensions.h: No such file or directory Configuring for: PHP Api Version: Zend Module Api No: Zend Extension Api No: Cannot find autoconf. Please check your autoconf installation and the $PHP_AUTOCONF environment variable. Then, rerun this script. ERROR: `phpize' failed Any tips on how to fix the errors and install OAuth succesfully? I'm on Mac OS X 10.6.3 Thanks!

    Read the article

  • How do I get information about the level to the player object?

    - by pangaea
    I have a design problem with my Player and Level class in my game. So below is a picture of the game. The problem is I don't want to move on the black space and only the white space. I know how to do this as all I need to do is get the check for the sf::Color::Black and I have methods to do this in the Level class. The problem is this piece of code void Game::input() { player.input(); } void Game::update() { (*level).update(); player.update(); } void Game::render() { (*level).render(); player.render(); } So as you there is a problem in that how do I get the map information from the Level class to the Player class. Now I was thinking if I made the Player position static and pass it into the Level as parameter in update I could do it. The problem is interaction. I don't know what to do. I could maybe make player go into the Level class. However, what if I want multiple levels? So I have big design problems that I'm trying to solve.

    Read the article

  • Blackberry message text is all white on white cannot read - Outlook 2007

    - by johnny
    Hi, I have a user that has Outlook 2007. When a certain person sends her an email from their blackberry the text is all white. If you copy all the text and place it in Word and change the font color you can see the email. The recipient is the only person that has the trouble with email from the blackberry device from this certain person. Everyone else can see the bb messages fine. Any ideas on what to check? I made sure the theme was set to none and that all fonts selected were installed (changed it to arial.) All other emails sent to the recipient are fine from everyone. The user that sends from the blackberry also has a PC. When he sends emails from the machine it looks fine for the troubled recipient. It is only when sending from the bb to that certain person that we get the white on white "invisible" text. Thank you for any help.

    Read the article

  • Now Available: Visual Studio 2010 Release Candidate Virtual Machines with Sample Data and Hands-on-L

    - by John Alexander
    From a message from Brian Keller: “Back in December we posted a set of virtual machines pre-configured with Visual Studio 2010 Beta 2, Visual Studio Team Foundation Server 2010 Beta 2, and 7 hands-on-labs. I am pleased to announce that today we have shipped an updated virtual machine using the Visual Studio 2010 Release Candidate bits, a brand new sample application, and 9 hands-on-labs. This VM is customer-ready and includes everything you need to learn and/or deliver demonstrations of many of my favorite application lifecycle management (ALM) capabilities in Visual Studio 2010. This VM is available in the virtualization platform of your choice (Hyper-V, Virtual PC 2007 SP1, and Windows [7] Virtual PC). Hyper-V is highly recommended because of the performance benefits and snapshotting capabilities. Tailspin Toys The sample application we are using in this virtual machine is a simple ASP.NET MVC 2 storefront called Tailspin Toys. Tailspin Toys sells model airplanes and relies on the application lifecycle management capabilities of Visual Studio 2010 to help them build, test, and maintain their storefront. Major kudos go to Dan Massey for building out this great application for us. Hands-on-Labs / Demo Scripts The 9 hands-on-labs / demo scripts which accompany this virtual machine cover several of the core capabilities of conducting application lifecycle management with Visual Studio 2010. Each document can be used by an individual in a hands-on-lab capacity, to learn how to perform a given set of tasks, or used by a presenter to deliver a demonstration or classroom-style training. Unlike the beta 2 release, 100% of these labs target Tailspin Toys to help ensure a consistent storytelling experience. Software quality: Authoring and Running Manual Tests using Microsoft Test Manager 2010 Introduction to Test Case Management with Microsoft Test Manager 2010 Introduction to Coded UI Tests with Visual Studio 2010 Ultimate Debugging with IntelliTrace using Visual Studio 2010 Ultimate Software architecture: Code Discovery using the architecture tools in Visual Studio 2010 Ultimate Understanding Class Coupling with Visual Studio 2010 Ultimate Using the Architecture Explore in Visual Studio 2010 Ultimate to Analyze Your Code Software Configuration Management: Planning your Projects with Team Foundation Server 2010 Branching and Merging Visualization with Team Foundation Server 2010 “ Check out Brian’s Post for more info including download instructions…

    Read the article

  • CentOS 5.8 server, installed Web Server vs yum install httpd

    - by Shiro
    I would like to know, what is the different between, the build-in CentOS server web server vs I install manual with yum install httpd I am new in CentOS. I just installed my CentOS 5.8 Server with Web Server checked. With some Google Search with LAMP installation, they listed with yum install httpd I had check inside /etc/init.d/ already had httpd I try to run yum install httpd http-devel It shows not yet installed. What will happen if I install yum version? What should I do? What is the best practise? Should I remove the Web Server install by default with CentOS? My goal is install LAMP (PHP v5.2.17).

    Read the article

  • Will my system fsck when I reboot?

    - by Tom Newton
    ...and how do I find out? Say I am about to reboot a server. I would like to minimize downtime, so thinking about wrapping reboot in an alias that says "hang on buddy, you're going to hit a fsck on boot". Next question.. what's the best way to say "lets do it next time?" set the last check date? I know tune2fs can set a bunch of parameters, but how would I get em?

    Read the article

  • Installing GeoIP on Mac?

    - by Richard
    I want to install the MaxMind GeoIP C library, so that I can use GeoIP as part of Django. I'm working on MacOS. I've downloaded GeoIP-1.4.8.tar.gz from here and run: ./configure make make check make install without any problems. I've also downloaded GeoIP.dat.gz and GeoLiteCity.dat.gz, unzipped them and put them in a local directory. Then I've set the following paths in my settings.py: GEOIP_PATH = 'path_to_dat_files' GEOIP_LIBRARY_PATH = 'path_to_c_files' Now I'm trying to run python manage.py migrate (which has a reference to GeoIP in a project I'm using), but I'm still getting this error: from django.contrib.gis.utils import GeoIP ImportError: cannot import name GeoIP Any ideas? Thanks.

    Read the article

  • DIVIDE vs division operator in #dax

    - by Marco Russo (SQLBI)
    Alberto Ferrari wrote an interesting article about DIVIDE performance in DAX. This new function has been introduced in SQL Server Analysis Services 2012 SP1, so it is available also in Excel 2013 (which still doesn’t have other features/fixes introduced by following Cumulative Updates…). The idea that instead of writing: IF ( Sales[Quantity] <> 0, Sales[Amount] / Sales[Quantity], BLANK () ) you can write: DIVIDE ( Sales[Amount], Sales[Quantity] ) There is a third optional argument in DIVIDE that defines the result in case the denominator (second argument) is zero, and by default its value is BLANK, so I omitted the third argument in my example. Using DIVIDE is very important, especially when you use a measure in MDX (for example in an Excel PivotTable) because it raise the chance that the non empty evaluation for the result is evaluated in bulk mode instead of cell-by-cell. However, from a DAX point of view, you might find it’s better to use the standard division operator removing the IF statement. I suggest you to read Alberto’s article, because you will find that an expression applying a filter using FILTER is faster than using CALCULATE, which is against any rule of thumb you might have read until now! Again, this is not always true, and depends on many conditions – trying to simplify, we might say that for a simple calculation, the query plan generated by FILTER could be more efficient – but, as usual, it depends, and 90% of the times using FILTER instead of CALCULATE produces slower performance. Do not take anything for granted, and always check the query plan when performance are your first issue!

    Read the article

  • Trouble installing Lagarith Lossless Codec (Codec Removal?)

    - by Xeoncross
    I had been using the Lagarith Lossless Codec with camstudio on Windows XP Pro SP3 for a little while before I switched to Ubuntu. Now I'm back trying to do something on windows and the codec is now missing. I tried to install it via the .exe and even manually with the .inf but it's not being listed by camstudio as installed. I'm wondering if one of the other codecs I installed might have messed with it and yet some of the files are still fine so whatever method the installers use to check for existing files before install is returning positive and not re-installing it. How can I get Lagarith to work again? Maybe I need to know how to uninstall codecs.

    Read the article

  • pure-ftpd: one readonly/non-deletable file in home directory

    - by Bram Schoenmakers
    Is there a way to have a file in the user's FTP home directory without the ability to modify/remove it from that directory over FTP? So the user has write permissions on his own home folder, thus the ability to remove files. An exception should be made for a single file, which has the same filename and contents for each account. The solution I'm thinking of right now to run a periodic script to check the presence of that file, and if not, put it back. But I wonder whether there's a better solution than this.

    Read the article

< Previous Page | 416 417 418 419 420 421 422 423 424 425 426 427  | Next Page >