Search Results

Search found 11268 results on 451 pages for 'shweta simply'.

Page 67/451 | < Previous Page | 63 64 65 66 67 68 69 70 71 72 73 74  | Next Page >

  • How can I set the default page for an https request?

    - by Rachel
    We have a website which has a Virtual Directory containing the secure portion of the website. If users come to http://www.mydomain.com, they should get directed to default.aspx of the main site, but if they go to https://www.mydomain.com, they should go to default.aspx of the virtual directory. The default page for the main site works fine, however I can't figure out how to set the default page for https traffic. Right now going to https://www.mydomain.com simply gives me a Page Not Found error. How can I set this up in IIS on Windows Server 2003?

    Read the article

  • Normals vs Normal maps

    - by KaiserJohaan
    I am using Assimp asset importer (http://assimp.sourceforge.net/lib_html/index.html) to parse 3d models. So far, I've simply pulled out the normal vectors which are defined for each vertex in my meshes. Yet I have also found various tutorials on normal maps... As I understand it for normal maps, the normal vectors are stored in each texel of a normal map, and you pull these out of the normal texture in the shader. Why is there two ways to get the normals, which one is considered best-practice and why?

    Read the article

  • Why are part-time jobs in programming an anomality?

    - by Mikle
    I've recently quit my full time developing job at mega-corp, and I decided that I'll look for a part time job. Since then I've talked to half a dozen potential employers, and every one of them had the same reaction when I said the magic words "part-time" - they all closed up and became suspicious. Now, I understand that it might just be me, so as control I asked every one of them what if I were willing to work full time, and they all said I would probably get an offer. My question is two fold: Why, as an employer, would you give up a competent, even great, developer, simply because he wants to work 3 days a week and not 5? How do I sell the story of part time job better? I usually just list my reasons which are that I prefer that balance currently in my life and that I want to work on my own projects, but it leaves them even more suspicious - am I going to start something myself and quit? Am I just lazy?

    Read the article

  • What's special about currying or partial application?

    - by Vigneshwaran
    I've been reading articles on Functional programming everyday and been trying to apply some practices as much as possible. But I don't understand what is unique in currying or partial application. Take this Groovy code as an example: def mul = { a, b -> a * b } def tripler1 = mul.curry(3) def tripler2 = { mul(3, it) } I do not understand what is the difference between tripler1 and tripler2. Aren't they both the same? The 'currying' is supported in pure or partial functional languages like Groovy, Scala, Haskell etc. But I can do the same thing (left-curry, right-curry, n-curry or partial application) by simply creating another named or anonymous function or closure that will forward the parameters to the original function (like tripler2) in most languages (even C.) Am I missing something here? There are places where I can use currying and partial application in my Grails application but I am hesitating to do so because I'm asking myself "How's that different?" Please enlighten me.

    Read the article

  • Referring to hardware/software in first-person? [closed]

    - by JYelton
    At my company, there is a habit for the engineers to refer to their respective hardware/firmware/software in the first-person as if the device they are responsible for is a manifestation of themselves. I'll give you an example: Hardware Engineer: "I don't receive the first byte, so I stay off." Software Engineer: "I'm sending you the first byte after the ack flag, so I thought you were getting it." Hardware Engineer: "No, you're not turning me on." It was this very example I overheard today that nearly had me giggling in fits. "You're not turning me on." Well, I should hope not! So, is it common practice for engineers to do this, or simply unprofessional? Any suggestions for changing this apparently bad habit?

    Read the article

  • Is software innovation still primarily North American and European? Why, and for how much longer?

    - by limist
    Since this site is read by a global audience of programmers, I want to know if people generally agree that the vast majority of software innovation - languages, OS, tools, methodologies, books, etc. - still originates from the USA, Canada, and the EU. I can think of a few exceptions, e.g. Nginx webserver from Russia and the Ruby language from Japan, but overwhelmingly, the software I use and encounter daily is from North America and the EU. Why? Is history and historical momentum (computing having started in USA and Europe) still driving the industry? And/or, is some nebulous (or real) cultural difference discouraging software innovation abroad? Or are those of us in the West simply ignorant of real software innovation going on in Asia, South America, Eastern Europe, etc.? When, if ever, might the centers of innovation move out of the West? Your experiences and opinions welcome, thanks!

    Read the article

  • How to shorten brain context switch delay when need to use new technology\framework?

    - by gasan
    The problem is when I have to deal with a new framework\library\language it completely slows my work process, at first it's kind of shock, you're sitting on your place about a day doing nothing surfing the net, because you simply can't do anything even read docs, then, on the second day I realize that I definitely should do something and starting read about it, then I realize that I don't understand it, then I'm reading until I got feeling that I should show some results immediately and then I'm writing the code quite fast and the job doesn't seem to be difficult. Then job is done and I won't probably return to that technology\framework for a month or a year or never at all. And I will almost certainly forget almost everything about it after a month. To illustrate by checkpoints I experience: shock, long studying times, work with the new tech briefly, never use it afterwards, then I completely forget it. So what would be the solution here?

    Read the article

  • How to price code reviews to encourage good behavior?

    - by Chris Clark
    I work for a company that has a hosted .net internet application with many clients. Those clients often want to write customizations for our application. We have APIs to hook into the app, but the customizations themselves are written in .net. This is a shared, secure hosting environment and we have to code review these customizations before we can deploy them in our datacenter to ensure that they don't degrade performance, crash our servers, or open any security vulnerabilities. We charge for these code reviews. The current pricing model is simply a function of the number of lines of code. I think this is a bad idea for a variety of reasons, but primarily because, if we are interested in verifying that the code works as expected, we should be incentivizing good, readable code, not compaction. I would like to propose a pricing model that incorporates some, or all of the following as inputs: Lines of code Cyclomatic complexity Avg function length # of functions Are there any other metrics I should incorporate, or other ideas for how we can reasonably create pricing for code reviews that encourages safe and understandable code?

    Read the article

  • Using JPA 2.0 with WebLogic Server 10.3.4 and Eclipse

    - by greg.stachnick
    Beginning in Oracle Enterprise Pack for Eclipse (OEPE) 11.1.1.6.1, we introduced a new feature for WebLogic Server configuration called Server Extensions. Similar in concept to project facets, Server Extensions allow us to install additional technologies, libraries, configurations, etc into an existing server runtime. WebLogic Server 10.3.4 introduces new support for Java Persistence 2.0, the new JEE 6 standard entity access. In order to start developing JPA 2.0 applications with WebLogic Server 10.3.4, a SmartUpdate patch must be applied to add and configure the EclipseLink libraries. More information on the manual EclipseLink installation and configuration can be found here. OEPE provides a Server Extension for JPA 2.0, making the addition and configuration of JPA 2.0 and EclipseLink much easier. When defining a new WebLogic Server 10.3.4 configuration, simply click the Install link for Java Persistence 2.0 and OEPE will take care of the WebLogic Server enablement for JPA 2.0.  

    Read the article

  • Parallel Installation of Ubuntu, Wubi and Windows 7?

    - by Aufwind
    I have Windows 7 installed on a Notebook. Because I wanted to experiment with Ubutnu without loosing my Windows 7 configuration I installed the Wubi Version of Ubuntu inside of Windows 7. Since things seam to work out well with Ubuntu and my harddisk space is very limited, I want to make Ubuntu my only operating system on this machine. So my question is: Is it possible, to install Ubuntu 11.04 additionally to my actual configuration (Wubi and Windows 7)? And if things turned out well with the native installation of Ubuntu, too, to delete Wubi and Windows 7 from my system afterwards? Why I am asking this is simply, that I am afraid, that if I just format my machine and made the whole harddisk available for the fresh Ubuntu installation, it could be possible, that it turns out, that one thing or an other would not work properly, like it did in Wubi. (Possible Network, VPN, and Firewall@work issues) In this case I'd like the possibility to switch back.

    Read the article

  • How To Colorize Black and White Vintage Photographs in Photoshop

    - by Eric Z Goodnight
    Ever wanted to add color to your old, vintage, or historical photographs? Load up some old pictures and see how color can be added quickly to any black and white photograph in this simple Photoshop how to. While many purists simply don’t like the look of colorized black and white photographs, the ability to add color to black and white images is as indispensible as it is simple. Read on to see just how easy it can be Latest Features How-To Geek ETC How To Create Your Own Custom ASCII Art from Any Image How To Process Camera Raw Without Paying for Adobe Photoshop How Do You Block Annoying Text Message (SMS) Spam? How to Use and Master the Notoriously Difficult Pen Tool in Photoshop HTG Explains: What Are the Differences Between All Those Audio Formats? How To Use Layer Masks and Vector Masks to Remove Complex Backgrounds in Photoshop Hack a Wireless Doorbell into a Snail Mail Indicator Enjoy Clutter-Free YouTube Video Viewing in Opera with CleanTube Bring Summer Back to Your Desktop with the LandscapeTheme for Chrome and Iron The Prospector – Home Dash Extension Creates a Whole New Browsing Experience in Firefox KinEmote Links Kinect to Windows Why Nobody Reads Web Site Privacy Policies [Infographic]

    Read the article

  • Community Branching

    - by Dane Morgridge
    As some may have noticed, I have taken a liking to Ruby (and Rails in particular) quite a bit recently. This last weekend I spoke at the NYC Code Camp on a comparison of ASP.NET and Rails as well as an intro to Entity Framework talk.  I am speaking at RubyNation in April and have submitted to other ruby conferences around the area and I am also doing a Rails and MongoDB talk at the Philly Code Camp in April. Before you start to think this is my "I'm leaving .NET post", which it isn't so I need to clarify. I am not, nor do I intend to any time in the near future plan on abandoning .NET.  I am simply branching out into another community based on a development technology that I very much enjoy.  If you look at my twitter bio, you will see that I am into Entity Framework, Ruby on Rails, C++ and ASP.NET MVC, and not necessarily in that order.  I know you're probably thinking to your self that I am crazy, which is probably true on several levels (especially the C++ part). I was actually crazy enough at the NYC Code Camp to show up wearing a Linux t-shirt, presenting with my MacBook Pro on Entity Framework, ASP.NET MVC and Rails. (I did get pelted in the head with candy by Rachel Appel for it though) At all of the code camps I am submitting to this year, i will be submitting sessions on likely all four topics, and some sessions will be a combination of 2 or more.  For example, my "ASP.NET MVC: A Gateway To Rails?" talk touches ASP.NET MVC, Entity Framework Code First and Rails. Simply put (and I talk about this in my MVC & Rails talk) is that learning and using Rails has made me a better ASP.NET MVC developer. Just one example of this is helper methods.  When I started working with ASP.NET MVC, I didn't really want to use helpers and preferred to just use standard html tags, especially where links were concerned.  It was just me being stubborn and not really seeing all of the benefit of the helpers.  To my defense, coming from WebForms, I wanted to be as bare metal as possible and it seemed at first like a lot of the helpers were an unnecessary abstraction. I took my first look at Rails back in v1 and didn't spend very much time with it so I dismissed it and went on my merry ASP.NET WebForms way.  Then I picked up ASP.NET MVC and grasped the MVC pattern itself much better. After this, I took another look at Rails and everything made sense.  I decided then to learn Rails. (I think it is important for developers to learn new languages and platforms regularly so it was a natural progression for me) I wanted to learn it the right way, so when I dug into code, everyone used helpers everywhere for pretty much everything possible. I took some time to dig in and found out how helpful they were and subsequently realized how awesome they were in ASP.NET MVC also and started using them. In short, I love Rails (and Ruby in general).  I also love ASP.NET MVC and Entity Framework and yes I still love C++.  I have varying degrees of love for them individually at any given moment and it is likely to shift based on the current project I am working on.  I know you're thinking it so before you ask the question. "Which do I use when?", I'm going to give the standard developer answer of: It depends.  There are a lot of factors that I am not going to even go into that would go into a decision.  The most basic question I would ask though is,  does this project depend on .NET?  If it does, then I'd say that ASP.NET MVC is probably going to be the more logical choice and I am going to leave it at that.  I am working on projects right now in both technologies and I don't see that changing anytime soon (one project even uses both). With all that being said, you'll find me at code camps, conferences and user groups presenting on .NET, Ruby or both, writing about .NET and Ruby and I will likely be blogging on both in the future.  I know of others that have successfully branched out to other communities and with any luck I'll be successful at it too. On a (sorta) side note, I read a post by Justin Etheredge the other day that pretty much sums up my feelings about Ruby as a language.  I highly recommend checking it out: What Is So Great About Ruby?

    Read the article

  • Adobe-Flash-plugin freezes when skipping ahead

    - by piedro
    I am using Firefox 7 and the new official Adobe Flash 64-bit plugin from the Ubuntu Partner repository. It's the adobe-flash package. It works nicely on the first look but when I try to skip ahead in any YouTube video (by clicking on the video progress bar) sound will go on but the video simply freezes at the position it was just before skipping. This is really annoying because I really don't want to watch every video full length. Now maybe there's cleanup work to do? I removed every Flash-related package from the package manager. Maybe it's Firefox? I couldn't find any hint. Maybe it's the video manager (display manager)? I am using KDE on an Ubuntu Natty install. I also enabled hardware acceleration with libdvpau1: sudo bash -c "echo 'EnableLinuxHWVideoDecode=1' >> /etc/adobe/mms.cfg" (this seems to have no effect at all ...) Ubuntu Natty 64-bit, Firefox 7, Flash 11.01 )

    Read the article

  • Is there a global "low resolution" filter for OpenGL?

    - by Ian Henry
    I'm trying to learn a little about OpenGL, so I'm making a simple 2D game (with OpenTK), and so far it's coming along well. I thought it would be fun to give it that, for lack of a better word, retropixelated look of games from the early nineties. I figured it would be an easy thing to do -- simply draw everything at half its normal size and scale up with no anti-aliasing. But I can't find any resources on how to do this. I can set the min/mag filters of my textures to nearest and that works fine for my sprites, but I'm using lots of primitives and I'd like the effect to apply to them as well. The one idea I had was to draw everything at half size, then somehow copy the render buffer to a texture, then render that texture full-size, but I don't know how to do that, and it seems like there must be a better way. Can anyone help me out?

    Read the article

  • Swap and hibernation

    - by maaartinus
    I saw a lot of recommendations claiming that for hibernation the swap partition/file must be at least as large as the main memory. This makes no sense to me. Lets assume I have 8 GB of main memory and 8 GB swap area and want to hibernate: case 1: I'm using 4 GB of virtual memory - 8 GB of swap is unnecessarily large. case 2: I'm using 8 GB of virtual memory - 8 GB of swap is just right. case 3: I'm using 12 GB of virtual memory - 8 GB of swap is too small. The outcome is: A swap area of size equal to the memory size is sufficient for hibernate IFF it doesn't get used for swapping at all. So what is the reason behind the claim that you need at least as much swap area as main memory for hibernate to work? I know that virtual memory gets used for caching too, and that the cache may be simply discarded, but what happens to hibernation if a program allocates 12 GB of virtual memory (given the above memory and swap sizes)?

    Read the article

  • Why is using C++ libraries so complicated?

    - by Pius
    First of all, I want to note I love C++ and I'm one of those people who thinks it is easier to code in C++ than Java. Except for one tiny thing: libraries. In Java you can simply add some jar to the build path and you're done. In C++ you usually have to set multiple paths for the header files and the library itself. In some cases, you even have to use special build flags. I have mainly used Visual Studio, Code Blocks and no IDE at all. All 3 options do not differ much when talking about using external libraries. I wonder why was there made no simpler alternative for this? Like having a special .zip file that has everything you need in one place so the IDE can do all the work for you setting up the build flags. Is there any technical barrier for this?

    Read the article

  • Should testers approve releases, or just report on tests?

    - by Ernest Friedman-Hill
    Does it make sense to give signoff authority to testers? Should a test team Just test features, issues, etc, and simply report on a pass/fail basis, leaving it up to others to act on those results, or Have authority to hold up releases themselves based on those results? In other words, should testers be required to actually sign off on releases? The testing team I'm working with feels that they do, and we're having an issue with this because of "testing scope creep" -- the refusal to approve releases is sometimes based on issues explicitly not addressed by the release in question.

    Read the article

  • Running an old version of some software

    - by Mark Oak
    I don't want to mingle in any backstory, but all that needs to be known is that I have a computer with Ubuntu on it and I am trying to install Windows 8 from an ISO. I am using the guide that can be found here which is a little more than four years old. Now, I've been able to accomplish everything up to Step 2, at which point I am stuck. I have downloaded the file found on that page, which can be found here, and have attempted to use it, as directed, quote; "right click the downloaded Unetbootin file, select Properties and on the "Permissions" tab, check the "Allow executing file as program" box. Then simply double click it and it should open." But, after having set checked the specified box and double clicking the file, nothing happens. Nothing is launched and nothing changes. I've been stuck here for several hours now, having failed to find a solution via Google.

    Read the article

  • How to clone a VirtualBox Disk

    - by [email protected]
     How to clone a VirtualBox DiskCopying the image of Virtual Disk (.vdi file) is a convenient way to duplicate the disk, in cases you want to avoid re-installing an operating system from scratch. However, simply copying the .vdi file into another location will make a verbatim copy of the virtual disk, including the UUID of the disk. If you try to add the copy in the Virtual Media Manager, you will get an error like this:In this case, you have to do is to clone the vdi disk: cd C:\Program Files\Sun\VirtualBox\C:\Program Files\Sun\VirtualBox>vboxmanage clonevdi G:\VMWARES\Database\11GR2onOEL5forVbox\11GR2_OEL5_32GB.vdi G:\VMWARES\Database\11GR2onOEL5forVbox\OEL5_32GB.vdi$ VBoxManage clonevdi Master.vdi Clone.vdiIn case you receive a error like this. It means that the disk is already a copy of other VirtualBox Disk.In that case you chould change the UUID before to clone the Disk.Follow the steps given here in order to do that.

    Read the article

  • How did programmers resolve their problems before the internet?

    - by 9a3eedi
    When programming, anytime I get stuck, perhaps with a compiler error that doesn't make sense, or from a GUI function that didn't do what I expected, I automatically google my problem, find someone else that faced the same thing, and read what's going on and why I'm getting the problem. Before the internet, how did people handle these situations? People used to read books and manuals more, I know. But books don't explain everything, like the odd compiler problem that you get sometimes, or nothing showing up on your screen despite you clearly writing correct OpenGL code. How did people cope when facing challenges? Did they simply "bash their head" on the wall till they figured it out? Is there something people used to do regularly on the side that gave them the ability to get themselves unstuck more easily? Were libraries/compilers much simpler back then? I've been asking this question because I sometimes feel guilty depending on Google so much when I'm pretty sure programmers before my time were more independent when it comes to facing these matters.

    Read the article

  • Pong Collision Help in C# w/ XNA

    - by Ramses Brown
    Edit: My goal is to have it function like this: Ball hits 1st Quarter = rebounds higher (aka Y++) Ball hits 2nd Quarter = rebounds higher (using random value) Ball hits 3rd Quarter = rebounds lower (using random value) Ball hits 4th Quarter = rebounds lower (aka Y--) I'm currently using Rectangle Collision for my collision detection, and it's worked. Now I wish to expand it. Instead of it simply detecting whether or not the paddle/ball intersect, I want to make it so that it can determine what section of the paddle gets hit. I wanted it in 4 parts, with each having a different reaction to impact. My first thought is to base it on the Ball's Y position compared to the Paddle's Y position. But since I want it in 4 parts, I don't know how to do that. So it's essentially be if (ball.Y > Paddle.Y) { PaddleSection1 == true; } Except modified so that instead of being top half/bottom half, it's 1st Quarter, etc.

    Read the article

  • A human-friendlier Samba name mangling

    - by Alex
    Most of our computers run Ubuntu, but two of them dual-boot into Windows, and when we have guests over, they typically also run Windows computers. Thus, in addition to using NFS, our file server (Ubuntu server) also runs Samba. And since we use Ubuntu mostly, we like to take advantage of its advantages over Windows, such as being able to use the characters \:*?"<>| in a file name. The problem, of course, is that Windows doesn't accept those characters in file names, and so Samba has to translate the file name into something more acceptable. The way it does this, however, I find to be obnoxious. The file name Episode 182 - Exorcist 2: The Heretic.mp4 for instance turns into E4Q82R~Y.MP4. This is a terrible "correction". Is there a way to make Samba's mangling a little more friendly to humans? Is possible to "correct" it to something like Episode 182 - Exorcist 2_ The Heretic.mp4 instead, where the illegal characters are simply substituted?

    Read the article

  • Strange GTK3 issue with Audacious and Evince (Document Viewer)

    - by Sman789
    I find that, in some GTK3 apps such as Audacious and Evince (Document Viewer), parts of the UI are pure black, which looks ugly and out-of-place. It is not intentional, as the screenshots on both websites show the app looking different. I only have this issue with the Adwaita GTK theme, but since it comes with Gnome 3 and is the main theme these apps are designed for, I'm guessing the problem is not simply the theme having some bug (though the Ubuntu package might). I have had this issue on two laptops now, running both Ubuntu 12.04 and Xubunu 12.10 with Gnome Shell and Adwaita. The screenshots below show the problem: Audcious's is visible in the search area around the text box, and behind the playlist tabs, and Evince's is visible in more or less the same position and around the document itself. Compare this from the Audacious homepage: With this image from my system: And Evince too:

    Read the article

  • How would you practice concurrency and multi-threading?

    - by Xavier Nodet
    I've been reading about concurrency, multi-threading, and how "the free lunch is over". But I've not yet had the possibility to use MT in my job. I'm thus looking for suggestions about what I could do to get some practice of CPU heavy MT through exercises or participation in some open-source projects. Thanks. Edit: I'm more interested in open-source projects that use MT for CPU-bound tasks, or simply algorithms that are interesting to implement using MT, rather than books or papers about the tools like threads, mutexes and locks...

    Read the article

  • Building a distributed system on Amazon Web Services

    - by Songo
    Would simply using AWS to build an application make this application a distributed system? For example if someone uses RDS for the database server, EC2 for the application itself and S3 for hosting user uploaded media, does that make it a distributed system? If not, then what should it be called and what is this application lacking for it to be distributed? Update Here is my take on the application to clarify my approach to building the system: The application I'm building is a social game for Facebook. I developed the application locally on a LAMP stack using Symfony2. For production I used an a single EC2 Micro instance for hosting the app itself, RDS for hosting my database, S3 for the user uploaded files and CloudFront for hosting static content. I know this may sound like a naive approach, so don't be shy to express your ideas.

    Read the article

< Previous Page | 63 64 65 66 67 68 69 70 71 72 73 74  | Next Page >