Search Results

Search found 54159 results on 2167 pages for 'system drawing color'.

Page 457/2167 | < Previous Page | 453 454 455 456 457 458 459 460 461 462 463 464  | Next Page >

  • Uniform not being applied to proper mesh

    - by HaMMeReD
    Ok, I got some code, and you select blocks on a grid. The selection works. I can modify the blocks to be raised when selected and the correct one shows. I set a color which I use in the shader. However, I am trying to change the color before rendering the geometry, and the last rendered geometry (in the sequence) is rendered light. However, to debug logic I decided to move the block up and make it white, in which case one block moves up and another block becomes white. I checked all my logic and it knows the correct one is selected and it is showing in, in the correct place and rendering it correctly. When there is only 1 it works properly. Video Of the bug in action, note how the highlighted and elevated blocks are not the same block, however the code for color and My Renderer is here (For the items being drawn) public void render(Renderer renderer) { mGrid.render(renderer, mGameState); for (Entity e:mGameEntities) { UnitTypes ut = UnitTypes.valueOf((String)e.getObject(D.UNIT_TYPE.ordinal())); if (ut == UnitTypes.Soldier) { renderer.testShader.begin(); renderer.testShader.setUniformMatrix("u_mvpMatrix",mEntityMatrix); renderer.texture_soldier.bind(0); Vector2 pos = (Vector2) e.getObject(D.COORDS.ordinal()); mEntityMatrix.set(renderer.mCamera.combined); if (mSelectedEntities.contains(e)) { mEntityMatrix.translate(pos.x, 1f, pos.y); renderer.testShader.setUniformf("v_color", 0.5f,0.5f,0.5f,1f); } else { mEntityMatrix.translate(pos.x, 0f, pos.y); renderer.testShader.setUniformf("v_color", 1f,1f,1f,1f); } mEntityMatrix.scale(0.2f, 0.2f, 0.2f); renderer.model_soldier.render(renderer.testShader,GL20.GL_TRIANGLES); renderer.testShader.end(); } else if (ut == UnitTypes.Enemy_Infiltrator) { renderer.testShader.begin(); renderer.testShader.setUniformMatrix("u_mvpMatrix",mEntityMatrix); renderer.testShader.setUniformf("v_color", 1.0f,1,1,1.0f); renderer.texture_enemy_infiltrator.bind(0); Vector2 pos = (Vector2) e.getObject(D.COORDS.ordinal()); mEntityMatrix.set(renderer.mCamera.combined); mEntityMatrix.translate(pos.x, 0f, pos.y); mEntityMatrix.scale(0.2f, 0.2f, 0.2f); renderer.model_enemy_infiltrator.render(renderer.testShader,GL20.GL_TRIANGLES); renderer.testShader.end(); } } }

    Read the article

  • Acer Extensa 5620 - Graphic unknown!

    - by Nycxzon
    I install Ubuntu 12.04 Beta 2, I found out the my graphic driver is "Unknown" and experience is "Standard". Please help me know how to install my graphic driver. as per my laptop specs: Mobile Intel GM965 Express Chipset with integrated 3D graphics, featuring Intel Graphics Media Accelerator (GMA) X3100 with up to 358/3845 MB of Intel Dynamic Video Memory Technology 4.0 (8 MB of dedicated system memory, up to 350/3765 MB of shared system memory), supporting Microsoft DirectX 9. Hope you can help me! Thanks in advance! Noob @ Ubuntu :) Updates! ______________ I found a solution! I found a site where I ask how to check the graphic of my system. by using terminal command "glxinfo" command ask me to install mesa-utils! and update. After that, My graphic driver is listed! Thanks for your continues support! :)

    Read the article

  • Component based design, but components rely on eatchother

    - by MintyAnt
    I've begun stabbing at a "Component Based" game system. Basically, each entity holds a list of components to update (and render) I inherit the "Component" class and break each game system into it. Examples: RenderComponent - Draws the entity MovementComponent - Moves the entity, deals with velocity and speed checks DamageComponent - Deals with how/if the entity gets damaged... So. My system has this: MovementComponent InputComponent Now maybe my design is off, but the InputComponent should say things like if (w key is down) add y speed to movement if (x key is down) Trigger primary attack This means that the InputComponent sort of relies on these other components. I have to do something alone the lines of: if (w key is down) { MovementComponent* entityMovement = mEntity->GetMovement(); if (entityMovement != NULL) add y speed to movement } which seems kinda crappy every update. Other options? Better design? Is this the best way? Thanks!

    Read the article

  • Can not use keyboard on unity

    - by ikhsan
    Dear Ubuntu Community, currently I am using Ubuntu 14.04, and few hours ago, an update notifier prompted to install an update. After update finished, it ask for system restart, I think there is some kernel update etc. The problem start after restart, I can type password when login, but after entering unity desktop, my keyboard become suddenly unusable, system doesn't respond to any key press, after few minutes, it lock the screen automatically, but still I can't type password to unlock the screen. I tried to logout (mouse is working properly), and login again, try starting onscreen keyboard, but still have no luck, system still doesn't respond to the key press. I tried to login in console, and keyboard working well, tried to install xfce, and keyboard also working properly, keyboard also working properly when login to unity as guest, it only not working when I login using my account. I also try to reset unity config via unity-tweak-tool --reset-unity , but still no luck any suggestion to resolve this?

    Read the article

  • Cloud Computing Forces Better Design Practices

    - by Herve Roggero
    Is cloud computing simply different than on premise development, or is cloud computing actually forcing you to create better applications than you normally would? In other words, is cloud computing merely imposing different design principles, or forcing better design principles?  A little while back I got into a discussion with a developer in which I was arguing that cloud computing, and specifically Windows Azure in his case, was forcing developers to adopt better design principles. His opinion was that cloud computing was not yielding better systems; just different systems. In this blog, I will argue that cloud computing does force developers to use better design practices, and hence better applications. So the first thing to define, of course, is the word “better”, in the context of application development. Looking at a few definitions online, better means “superior quality”. As it relates to this discussion then, I stipulate that cloud computing can yield higher quality applications in terms of scalability, everything else being equal. Before going further I need to also outline the difference between performance and scalability. Performance and scalability are two related concepts, but they don’t mean the same thing. Scalability is the measure of system performance given various loads. So when developers design for performance, they usually give higher priority to a given load and tend to optimize for the given load. When developers design for scalability, the actual performance at a given load is not as important; the ability to ensure reasonable performance regardless of the load becomes the objective. This can lead to very different design choices. For example, if your objective is to obtains the fastest response time possible for a service you are building, you may choose the implement a TCP connection that never closes until the client chooses to close the connection (in other words, a tightly coupled service from a connectivity standpoint), and on which a connection session is established for faster processing on the next request (like SQL Server or other database systems for example). If you objective is to scale, you may implement a service that answers to requests without keeping session state, so that server resources are released as quickly as possible, like a REST service for example. This alternate design would likely have a slower response time than the TCP service for any given load, but would continue to function at very large loads because of its inherently loosely coupled design. An example of a REST service is the NO-SQL implementation in the Microsoft cloud called Azure Tables. Now, back to cloud computing… Cloud computing is designed to help you scale your applications, specifically when you use Platform as a Service (PaaS) offerings. However it’s not automatic. You can design a tightly-coupled TCP service as discussed above, and as you can imagine, it probably won’t scale even if you place the service in the cloud because it isn’t using a connection pattern that will allow it to scale [note: I am not implying that all TCP systems do not scale; I am just illustrating the scalability concepts with an imaginary TCP service that isn’t designed to scale for the purpose of this discussion]. The other service, using REST, will have a better chance to scale because, by design, it minimizes resource consumption for individual requests and doesn’t tie a client connection to a specific endpoint (which means you can easily deploy this service to hundreds of machines without much trouble, as long as your pockets are deep enough). The TCP and REST services discussed above are both valid designs; the TCP service is faster and the REST service scales better. So is it fair to say that one service is fundamentally better than the other? No; not unless you need to scale. And if you don’t need to scale, then you don’t need the cloud in the first place. However, it is interesting to note that if you do need to scale, then a loosely coupled system becomes a better design because it can almost always scale better than a tightly-coupled system. And because most applications grow overtime, with an increasing user base, new functional requirements, increased data and so forth, most applications eventually do need to scale. So in my humble opinion, I conclude that a loosely coupled system is not just different than a tightly coupled system; it is a better design, because it will stand the test of time. And in my book, if a system stands the test of time better than another, it is of superior quality. Because cloud computing demands loosely coupled systems so that its underlying service architecture can be leveraged, developers ultimately have no choice but to design loosely coupled systems for the cloud. And because loosely coupled systems are better… … the cloud forces better design practices. My 2 cents.

    Read the article

  • "kde-config not found" error while installing kstars from source

    - by tachyons
    I am trying to install kstars from source ,but I got the following error while configuring ./configure checking build system type... i686-pc-linux-gnu checking host system type... i686-pc-linux-gnu checking target system type... i686-pc-linux-gnu checking for a BSD-compatible install... /usr/bin/install -c checking for -p flag to install... yes checking whether build environment is sane... yes checking for gawk... no checking for mawk... mawk checking whether make sets $(MAKE)... yes checking for kde-config... not found configure: error: The important program kde-config was not found! Please check whether you installed KDE correctly. What does it mean ,I already installed kde in my computer

    Read the article

  • Are these non-standard applications of rendering practical in games?

    - by maul
    I've recently got into 3D and I came up with a few different "tricky" rendering techniques. Unfortunately I don't have the time to work on this myself, but I'd like to know if these are known methods and if they can be used in practice. Hybrid rendering Now I know that ray-tracing is still not fast enough for real-time rendering, at least on home computers. I also know that hybrid rendering (a combination of rasterization and ray-tracing) is a well known theory. However I had the following idea: one could separate a scene into "important" and "not important" objects. First you render the "not important" objects using traditional rasterization. In this pass you also render the "important" objects using a special shader that simply marks these parts on the image using a special color, or some stencil/depth buffer trickery. Then in the second pass you read back the results of the first pass and start ray tracing, but only from the pixels that were marked by the "important" object's shader. This would allow you to only ray-trace exactly what you need to. Could this be fast enough for real-time effects? Rendered physics I'm specifically talking about bullet physics - intersection of a very small object (point/bullet) that travels across a straight line with other, relatively slow-moving, fairly constant objects. More specifically: hit detection. My idea is that you could render the scene from the point of view of the gun (or the bullet). Every object in the scene would draw a different color. You only need to render a 1x1 pixel window - the center of the screen (again, from the gun's point of view). Then you simply check that central pixel and the color tells you what you hit. This is pixel-perfect hit detection based on the graphical representation of objects, which is not common in games. Afaik traditional OpenGL "picking" is a similar method. This could be extended in a few ways: For larger (non-bullet) objects you render a larger portion of the screen. If you put a special-colored plane in the middle of the scene (exactly where the bullet will be after the current frame) you get a method that works as the traditional slow-moving iterative physics test as well. You could simulate objects that the bullet can pass through (with decreased velocity) using alpha blending or some similar trick. So are these techniques in use anywhere, and/or are they practical at all?

    Read the article

  • Passing functions into other functions as parameters, bad practice?

    - by BlueHat
    We've been in the process of changing how our AS3 application talks to our back end and we're in the process of implementing a REST system to replace our old one. Sadly the developer who started the work is now on long term sick leave and it's been handed over to me. I've been working with it for the past week or so now and I understand the system, but there's one thing that's been worrying me. There seems to be a lot of passing of functions into functions. For example our class that makes the call to our servers takes in a function that it will then call and pass an object to when the process is complete and errors have been handled etc. It's giving me that "bad feeling" where I feel like it's horrible practice and I can think of some reasons why but I want some confirmation before I propose a re-work to system. I was wondering if anyone had any experience with this possible problem?

    Read the article

  • Getting the front buffer into a gfx mem surface (Dx9)

    - by lapin
    I'm using DirectX 9 to acquire the frontbuffer. There are a couple of ways I know of to get at the front buffer: GetRenderTargetData() GetFrontBufferData() The MSDN page on both of these API calls state that the data is copied from device memory to system memory. I'd like to copy the front buffer surface directly to another graphics memory surface, as I have other manipulations to perform on the acquired surface before returning it to system memory. I'm creating a D3DUSAGE_DYNAMIC texture (gfx mem texture) and calling GetFrontBufferData() to write the front buffer to my textures surface0. Is this valid? Will the operation remain in gfx memory, or will it need to move to system memory and then back to graphics memory? If this is the case, is what I'm trying to achieve possible?

    Read the article

  • No dual boot menu

    - by Christian Galo
    I formatted all of my disk and installed Ubuntu on my computer. I immediately partitioned, from an Ubuntu live CD, my hard drive, creating an NTFS partition for for Windows. After successfully doing so, I went on to install windows 8.1. After I installed Windows 8 in the new partition and turned off my PC and turned it on again the option to chose which Operating system I wanted to use didn't appear, loading Windows like if Ubuntu didn't exist. How can I have the option to chose which operating system I want to run or at least from which partition to boot from when I start my computer? EVERYTHING IS OKAY WITH MY OPERATING SYSTEM The only thing I need is for ubuntu to appear as an option on the boot menus

    Read the article

  • How to conduct A/B split testing with AdSense?

    - by None
    Ok so I have decided to A/B test my AdSense ads. I have run a few tests, but I don't know what conclusion to draw and how to keep track of things. Some specific questions: If I have 2 test units, 1 wins. I test that with a new and so on. How do I find if say the fifth one did better than the first one? How do I keep track of things? Do I let the variables independent of each other, because they certainly are not. In real life, font size can affect CTR even if the colors are different. I can test blue color with red color, and then test Arial font with Georgia, but how do I know which combination is the best? This would result in way too many test units. I tried Googling a lot, but I could not find answers to these questions.

    Read the article

  • Which code module should map physical keys to abstract keys?

    - by Paul Manta
    How do you bridge the gap between the library's low-level event system and your engine's high-level event system? (I'm not necessarily talking about key events, but also about quit events.) At the top level of my event system, I send out KeyPressedEvents, KeyRelesedEvents and others of this kind. These high-level events only contain the abstract values of the keys (they don't say that Space way pressed, but that the JumpKey was pressed, for example). Whose responsibility should it be to map the "JumpKey" to an actual key on the keyboard?

    Read the article

  • Can realtek 8192cu usb wireless card be used in ubuntu 12.04 with kernel 3.2.0?

    - by waterloo2005
    I do like the post RTL8188CUS Wireless USB Dongle doesn't work unless I disable wireless security . But when I plug my 8192cu usb wireless card, my computer screen is off. At that time I even can not use Alt SysRq k or Alt SysRq + reisub. I compile the latest driver of 8192cu on ubuntu 12.04 with kernel 3.2.0-34. In RealTek site, I download 8192cu drive which is for Linux Kernel 2.6.18~2.6.38 and Kernel 3.0.8. But now in Ubuntu12.04 my kernel is 3.2.0-34. Every time I plug the usb 8192cu wireless card my system halts. Now I try to blacklist both the system's rtl8192cu driver and the new 8192cu driver I compile, but the system still halts when I plug the usb driver. What about you ? Thanks!

    Read the article

  • 12.04 Screen goes OFF and ON and hangs

    - by SKC
    i have installed ubuntu 12.04. while i'm using the screen goes off, comes back on (during which time, the system is not responsive) and this process repeats one more time in immediate succession after which the computer is responsive again. sometimes the system hangs indefinitely. sometimes the windows go blank. only the title bars are seen. i have tried re-installing the OS three times with no improvements. i have tried both 32 bit as well as 64 bit. i have no idea what the problem is and have tried hard to pin point that. i never had this prob till 11.10 my system config is as follow processor - 2nd gen intel core i7 (2660k) graphics - intel onboard HD graphics 3000 (no external graphics card) RAM - 4GB none of my friends seem to experience this problem. it really gets annoying as it happens really often. please advice

    Read the article

  • Best way to determine surface normal for a group of pixels?

    - by Paul Renton
    One of my current endeavors is creating a 2D destructible terrain engine for iOS Cocos2D (See https://github.com/crebstar/PWNDestructibleTerrain ). It is in an infant stages no doubt, but I have made significant progress since starting a couple weeks ago. However, I have run into a bit of a performance road block with calculating surface normals. Note: For my destructible terrain engine, an alpha of 0 is considered to not be solid ground. The method posted below works just great given small rectangles, such as n < 30. Anything above 30 causes a dip in the frame rate. If you approach 100x100 then you might as well read a book while the sprite attempts to traverse the terrain. At the moment this is the best I can come up with for altering the angle on a sprite as it roams across terrain (to get the angle for a sprite's orientation just take dot product of 100 * normal * (1,0) vector). -(CGPoint)getAverageSurfaceNormalAt:(CGPoint)pt withRect:(CGRect)area { float avgX = 0; float avgY = 0; ccColor4B color = ccc4(0, 0, 0, 0); CGPoint normal; float len; for (int w = area.size.width; w >= -area.size.width; w--) { for (int h = area.size.height; h >= -area.size.height; h--) { CGPoint pixPt = ccp(w + pt.x, h + pt.y); if ([self pixelAt:pixPt colorCache:&color]) { if (color.a != 0) { avgX -= w; avgY -= h; } // end inner if } // end outer if } // end inner for } // end outer for len = sqrtf(avgX * avgX + avgY * avgY); if (len == 0) { normal = ccp(avgX, avgY); } else { normal = ccp(avgX/len, avgY/len); } // end if return normal; } // end get My problem is I have sprites that require larger rectangles in order for their movement to look realistic. I considered doing a cache of all surface normals, but this lead to issues of knowing when to recalculate the surface normals and these calculations also being quite expensive (also how large should the blocks be?). Another smaller issue is I don't know how to properly treat the case when length is = 0. So I am stuck... Any advice from the community would be greatly appreciated! Is my method the best possible one? Or should I rethink the algorithm? I am new to game development and always looking to learn new tips and tricks.

    Read the article

  • Reboot only shuts down and doesn't actually boot again

    - by PherricOxide
    I'm running a fresh install of Ubuntu 12.04 Server on an abmx rack mount server. When I attempt to reboot with sudo shutdown -r now, the machine just shuts down and doesn't come back up without me manually pressing the power button. The output of last -x, runlevel (to lvl 2) 3.2.0-29-generic Wed Oct 31 14:32 - 14:37 (00:05) reboot system boot 3.2.0-29-generic Wed Oct 31 14:32 - 14:37 (00:05) shutdown system down 3.2.0-29-generic Wed Oct 31 14:30 - 14:32 (00:02) runlevel (to lvl 6) 3.2.0-29-generic Wed Oct 31 14:30 - 14:30 (00:00) This appears to show that the system rebooted, but it went dark (no power lights, BIOS, etc) and I had to go press the power button in order to make it boot back up. The machine does have some sort of Intel Boot Agent that usually appears before the BIOS, I'm wondering if it could be causing this. I'm not sure what information is useful for debugging this, but I put the output of lshw in http://pastebin.com/mBy72kTQ

    Read the article

  • Styling Windows Phone Silverlight Applications

    - by Tim Murphy
    If you have not developed with styles in Silverlight/XAML then it can be challenging and resources can be sparse depending on how deep you get.  One thing that you need to understand is what level you can apply styles and how much they can cascade.  What I am finding is that this doesn’t go to the level that we are used to in HTML and CSS. While styles can be defined at a page level if you want to share styles throughout your application they should be defined in the App.xaml file.  This is of course analogous to placing a style in your HTML file versus an external CSS file.  This is the type of style I will concentrate on in this post. The first thing to look it how styles associate to elements.  TargetType defines the object type that your style will apply to.  In the example below the style is targeting the TextBlock object type. <Style x:Key="TextBlockSmallGray" TargetType="TextBlock"> Next we use a Setter which allows you to apply values for specific attributes of the target object type.  The setters can be a simple value or complex.  The first example here is simply applying a color to the background property of the target. <Setter Property="Background" Value="White"/> The second setter example here is for the same property, but we are applying a the definition of a LinearGradientBrush. <Setter Property="Background"> <Setter.Value> <LinearGradientBrush> <GradientStop Offset="0" Color="Black"/> <GradientStop Offset="1" Color="White"/> </LinearGradientBrush> </Setter.Value> </Setter> The last thing I want to cover here is that you can leverage the system styles and then override or extend them.  The BasedOn attribute of the Style tag allows this sort of inheritance.  In the example below I am going to start with the PhoneTextTitleStyle and then override properties as needed. <Style x:Key="TextBlockTitle" BasedOn="{StaticResource PhoneTextTitle1Style}" TargetType="TextBlock"> So now that we have our styles defined applying it is fairly straight forward.  Add the style name as a static resource to the style property of the element in your page and off you go. <Grid x:Name="LayoutRoot" Style="{StaticResource PageGridStyle}"> So this is one step in creating consistency in your application’s look.  In future posts I will dig a little deeper. del.icio.us Tags: windows phone 7,mobile development,windows phone 7 development,.NET,software development,design,UX

    Read the article

  • vlc 1.1.9 not working properly

    - by jaggib
    I have installed vlc 1.1.9 on wubi installed Ubuntu 11.04 using Ubuntu Software Center. Now when i tried playing videos in vlc (any format) full screen mode doesnt show controls and usually doesnt gets me out to window mode. it sometimes crashes to login screen. I tried the video output to 'X11 output mode' and 'XVideo output (XCB)' but the above problem persists and also bring another problem. full screen mode doesnt responds always when it does, it shows the video over the desktop instead of inside the player. the only way to get ubuntu to function normally is to restart the system. I tried with not using the 'Embed video in interface' but still the same problem. vlc runs perfectly well in windows. How can i make vlc function properly on my system or i need to install another player? My system config is: Graphics: VIA/S3G UniChrome Pro IGP Processor: AMD Sempron(tm) Processor 2800+ RAM: 1.5GB Motherboard Name: MS-7181 (MSI)

    Read the article

  • XNA GameTime TotalGameTime slower than real time

    - by robasaurus
    I have set-up an empty test project consisting of a System.Diagnostics.Stopwatch and this in the draw method: spriteBatch.DrawString(font, gameTime.TotalGameTime.TotalSeconds.ToString(), new Vector2(100, 100), Color.White); spriteBatch.DrawString(font, stopwatch.Elapsed.TotalSeconds.ToString(), new Vector2(100, 200), Color.White); The GameTime.TotalGameTime displayed is slower than the stop watch (by about 5 seconds per minute) even though GameTime.IsRunningSlowly is always false, why is this? The reason this is an issue is because I have a server which uses stopwatch and it is faster than my client game. For instance my client notifies the server it has dropped a mine which explodes in one minute. Because the stopwatch is faster the server state explodes the mine before the client and they are out of sync. I don't want to have to notify the client when the server explodes it as this would use unnecessary bandwidth.

    Read the article

  • Announcement: Oracle SuperCluster T5-8

    - by uwes
    Oracle's Fastest Engineered System On 27th of June we are announcing Oracle SuperCluster T5-8, Oracle’s fastest engineered system. Combining powerful virtualization and unique Exadata and Exalogic optimizations, SuperCluster is optimized to run both database and enterprise applications, and is ideal for consolidation and private cloud. SuperCluster is a complete system integrating SPARC T5-8 servers, Exadata Storage Servers, ZFS Storage Appliance, InfiniBand network and software, delivering extreme performance, no single point of failure, and highest efficiency while reducing risks and costs. Leverage Oracle SuperCluster T5-8 for IBM and HP competitive displacements, upgrading existing data centers, or new customer deployments. Please read the Product Bulletin on Oracle HW TRC for more details. (If you are not registered on Oracle HW TRC, click here ... and follow the instructions..) For More Information Go To: Oracle SuperCluster T5-8 oracle.com OTN

    Read the article

  • Using packages (gems, eggs, etc.) to create decoupled architectures

    - by Juan Carlos Coto
    The main issue Seeing the good support most modern programming platforms have for package management (think gem, npm, pip, etc), does it make sense to design an application or system be composed of internally developed packages, so as to promote and create a loosely coupled architecture? Example An example of this would be to create packages for database access, as well as for authentication and other components of the system. These, of course, use external packages as well. Then, your system imports and uses these packages - instead of including their code within its own code base. Considerations To me, it seems that this would promote code decoupling and help maintainability, almost in a Web-based-vs.-desktop-application kind of way (updates are applied almost automatically, single code base for single functionality, etc.). Does this seem like a rational and sane design concept? Is this actually used as a standard way of structuring applications today? Thanks very much!

    Read the article

  • Most appropriate diagram for GUI button design?

    - by JustADude
    What is the most appropriate diagram for GUI button design? Specifically, I have numerous buttons that will be changing color based on state information from operator input and input from various subcomponents and sensors. I would like to use UML or some other type of design diagrams to be able to capture the color transition. Some folks have suggested sequence diagrams, but I haven't been able to find any good examples that show how to incorporate this design artifact. I would really like to head in this direction to help developers. Thanks a ton for feedback and insights.

    Read the article

  • Is it easier to develop from scratch or not? [closed]

    - by Gnijuohz
    I am currently reading the book computer science: an overview by J. Glenn Brookshear.And In chapter 7,there is one passage as follows: In fact, it is often within this phase(modification) that a piece of software is discarded under the pretense (too often true) that it is easier to develop a new system from scratch than to modify the existing package successfully. When I read this,an article by Joel occurred to me which mentions how Mozilla shouldn't have written its browser from scratch.(The article is here) So,Is it mostly true that it's easier to develop a new system from scratch than to modify the existing one?Or it's closely related to the complexity of the system?

    Read the article

  • Highlighting new rows in ADF Table

    - by Sireesha Pinninti
    About This article explains how to hightlight newly inserted rows in an ADF Table without writing any extra java/javascript code.IntroductionSometimes we may wish to give more clarification to the end user by differentiating between newly inserted rows and the existing rows(i.e the rows from DB) in a table by highlighting new rows in different color as in the figure shown below. SolutionWe can achieve the same by giving following EL to inlineStyle property of every column inside af:table: #{row.row.entities[0].entityState == 0?'background-color:#307D7E;':''}ExplanationHere is the explanation for row.row.entities[0].entityState given inside EL which returns the state of the row(i.e, New, Modified, Unmodified, Initialized etc.)row - Refers to a tree node binding(instance of FacesCtrlHierNodeBinding) at runtimerow.row - Refers to an instance of row that the tree node is based onrow.row.entities[0] - Gets the Entity row at zeroth index. In most of the cases, the table will be based on single entity. If your table is based on multiple entities then the index needs to be given accordingly.row.row.entities[0].entityState - Gets Entity Object's current Entity-state in the transaction.(0 - New, Modified - 2, Unmodified - 1, Initialized - -1,  etc.,)

    Read the article

  • Why does unity obj import flip my x coordinate?

    - by milkplus
    When I import my wavefront obj model into unity and then draw lines over it with the same coordinates in the obj file, the x coordinate is negated. I don't see any option in the importer that might be doing that. And I'm using the same localToWorldMatrix and the same coordinate data in the .obj file. Hmmm GL.PushMatrix(); GL.MultMatrix(transform.localToWorldMatrix); CreateMaterial(); lineMaterial.SetPass(0); GL.Color(new Color(0, 1, 0)); GL.Begin(GL.LINES); GL.Vertex(p1); GL.Vertex(p2); GL.Vertex(p2); GL.Vertex(p3); //... GL.End(); GL.PopMatrix();

    Read the article

< Previous Page | 453 454 455 456 457 458 459 460 461 462 463 464  | Next Page >